Equal numbers of neuronal and nonneuronal cells make the human brain an isometrically scaled-up primate brain.

Instituto de Ciências Biomédicas, Universidade Federal do Rio de Janeiro, Cidade Universitária, Rio de Janeiro, Brazil.
The Journal of Comparative Neurology (Impact Factor: 3.51). 05/2009; 513(5):532-41. DOI: 10.1002/cne.21974
Source: PubMed

ABSTRACT The human brain is often considered to be the most cognitively capable among mammalian brains and to be much larger than expected for a mammal of our body size. Although the number of neurons is generally assumed to be a determinant of computational power, and despite the widespread quotes that the human brain contains 100 billion neurons and ten times more glial cells, the absolute number of neurons and glial cells in the human brain remains unknown. Here we determine these numbers by using the isotropic fractionator and compare them with the expected values for a human-sized primate. We find that the adult male human brain contains on average 86.1 +/- 8.1 billion NeuN-positive cells ("neurons") and 84.6 +/- 9.8 billion NeuN-negative ("nonneuronal") cells. With only 19% of all neurons located in the cerebral cortex, greater cortical size (representing 82% of total brain mass) in humans compared with other primates does not reflect an increased relative number of cortical neurons. The ratios between glial cells and neurons in the human brain structures are similar to those found in other primates, and their numbers of cells match those expected for a primate of human proportions. These findings challenge the common view that humans stand out from other primates in their brain composition and indicate that, with regard to numbers of neuronal and nonneuronal cells, the human brain is an isometrically scaled-up primate brain.

  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In broad strokes this chapter examines the roots of recurrence quantification analysis (RQA) and looks forward into its future. Much has already been written about the theory and application of recurrence methodology to real-world and complex systems. The overarching purpose of this present contribution is to introduce recurrence quantification as a unifying tool that has the timely potential of bringing together numerous and diverse fields of science under a common rubric of dynamical systems. The simple idea is that all scientists are studying "linguistic" systems that possess unique "vocabularies" and varied "grammars." But RQA can cut through these differences and encourage researchers in different research fields and scattered geographical locals to start talking seriously about the similarities of otherwise apparently diverse systems. 1. Context The concept of recurrence has a long mathematical history that can be traced back at least to Poincare (1880) of the late 19 th century and more recently to Feller (1950) of the mid 20 th century. Poincare may be considered the "father" of nonlinear dynamics, whereas Feller was a mathematician focusing on probability theory. As we will see, recurrence analysis is actually a statistical tool that can "diagnose" any moving system, but we are getting ahead of ourselves. Systems can be considered as recurrent if dynamic trajectories describing those systems return exactly to (periodic) or nearby (chaotic) their starting states. According this definition, recurrent systems possess periodicities or quasi-periodicities which keep such systems intact and functioning. Any active system with strong recurrences can be considered to be in some type of steady state or quasi-steady state. However, any system exhibiting weak recurrences is either transitioning to another steady state or is moving toward dissolution (cessation). And theoretically, systems possessing no or at best unstructured recurrences are no systems at all in the classical sense, and are better understood as random processes (which nonetheless may have high-dimensional structuring that are undetermined as yet). In the second half of the 20 th century three key mathematical breakthroughs (greatly aided by the rise of digital computers) have preceded the inception of modern recurrence quantification analysis (RQA). First, in 1963 Lorenz (1963) (13,559 citations) while modeling weather systems found that small changes in initial conditions in his three-variable system led to unpredictable dynamic outcomes. This modern finding was reminiscent of Newton's three-body problem which Poincare proved had no analytical solution. Thus was born the famous Lorenz strange attractor which digital and analog computers could easily display. Besides displaying sensitivity to initial conditions, strange attractors have return trajectories that are very close but non-exact. Second, in 1981 Takens (1981) (8,206 citations) published results from his work on strange attractors in fluid turbulence. His huge contribution was the mathematical proof that dynamical systems in N dimensions could be reproduced typologically
    The CISB scientific activity: Recent and seminal achievements., Rome, Italy; 05/2014
  • Source
    Manganese in Health and Disease, Edited by Lucio Costa, Michael Aschner, 08/2014: chapter 22: pages 540-573; Royal Society of Chemistry.
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In recent years, it became clear that a better understanding of the interactions among the main elements involved in the cancer network is necessary for the treatment of cancer and the suppression of cancer growth. In this work we propose a system of coupled differential equations that model brain tumour under treatment by chemotherapy, which considers interactions among the glial cells, the glioma, the neurons, and the chemotherapeutic agents. We study the conditions for the glioma growth to be eliminated, and identify values of the parameters for which the inhibition of the glioma growth is obtained with a minimal loss of healthy cells. Copyright © 2015. Published by Elsevier Ltd.
    Journal of Theoretical Biology 01/2015; 368. · 2.35 Impact Factor

Full-text (2 Sources)

Available from
May 27, 2014