「entropy」の共起表現一覧(1語右で並び替え)
該当件数 : 242件
am was still in possession of their own flag; | Entropy, a game of strategy in which player teams att |
Any structure implicit in the finite | entropy against a quantum description could then be t |
The normal distribution N(μ,σ2) has maximum | entropy among all real-valued distributions with spec |
and Bruce J. West, “Multiresolution diffusion | entropy analysis of time series: an application to bi |
A quantum black hole only has finite | entropy and therefore presumably exists in one of a l |
arity properties of a Gaussian process to its | entropy and covariance structure. |
er Strassen for making the connection between | entropy and regularity. |
that cosmological event horizons also have an | entropy and temperature. |
eries Wonders of the Universe to help explain | Entropy and its effect on time. |
hropomorphic personification of the forces of | entropy and chaos, the counterpart of the White Guard |
entropy to demonstrate the similarity between | entropy and information. |
e Planck, "the logarithmic connection between | entropy and probability was first stated by L. Boltzm |
temperature and, by the no hair theorem, zero | entropy, and the laws of black hole mechanics remain |
irreversible circuit, the information becomes | entropy and an associated amount of energy is dissipa |
areas of physics of information; information, | entropy, and complexity; quantum information theory; |
y discover that Logopolis is being overrun by | entropy and turning to dust. |
echnique, leading to the principle of maximum | entropy, and the MaxEnt view of thermodynamics. |
ng and expands for some time, with increasing | entropy and a thermodynamic arrow of time pointing in |
By using the following basic equations for | entropy and Helmholtz free energy, we can model the d |
that of Ludwig Boltzmann's 1870s theories of | entropy and order and disorder. |
pers on various topics such as randomness and | entropy, and teaching geometry to children. |
first mathematical version of the concept of | entropy, and also gave it its name. |
rime, also known as "the Fallen"; overseer of | entropy and Warrior of Darkness, now a herald of Unic |
based on the thermodynamic definition of the | entropy, and require extremely careful calorimetry. |
fated plea to vengeance demon Anyanka; and in | Entropy, Anya's unsuccessful attempts to enact her ow |
relative entropies subtracted from the total | entropy are 0. |
ests that the idea of thermodynamic state and | entropy are somewhat subjective. |
tion, Frank L. Lambert argued that portraying | entropy as "disorder" is confusing and should be aban |
nally, 20th century textbooks have introduced | entropy as order and disorder so that it provides "a |
Then, in 2002 “ | Entropy Is Simple , Qualitatively” described the spon |
ixed by matching the semiclassical black hole | entropy, as calculated by Stephen Hawking, and the co |
hing arbitrarily high temperature and maximal | entropy as the end of the collapse approaches. |
is possible (in a thermal context) to regard | entropy as an indicator or measure of the effectivene |
ion (the first of which sometimes called free | entropy), as cited by American engineer Willard Gibbs |
n associated with Frank L. Lambert describing | entropy as energy dispersal describes entropy as meas |
The conformational | entropy associated with a particular conformation is |
The "disorder" and hence the | entropy associated with the change in the atomic alig |
rature and the area of the event horizon with | entropy, at least up to some multiplicative constants |
Entropy Automatic (self-released, 2004) | |
During steady-state continuous operation, an | entropy balance applied to an open system accounts fo |
These have residual | entropy, because the atom-by-atom microscopic structu |
In information theory, the cross | entropy between two probability distributions measure |
, the holographic principle and the covariant | entropy bound of quantum gravity, and can be derived |
olographic bound, also known as the covariant | entropy bound. |
define a life form as an entity that reduces | entropy by self-executing the minimum set of physical |
hy even though Hawking radiation has non-zero | entropy, calculations so far have been unable to rela |
inal work), from which the correct Bekenstein | entropy can then be deduced. |
Their black hole | entropy can often be calculated in string theory, muc |
ity can be roughly interpreted as saying that | entropy can only be canceled by an equal amount of en |
e qualitative relation of energy dispersal to | entropy change can be so inextricably obscured that i |
nd to be 6008 joules at 273 K. Therefore, the | entropy change per mole is , or 22 J/K. |
Entropy change is the quantitative measure of that ki | |
inative model that extends a standard maximum | entropy classifier by assuming that the unknown value |
s of features and classes, while in a maximum | entropy classifier the weights, which are typically m |
ogit model, it is commonly known as a maximum | entropy classifier, or MaxEnt model for short. |
e classifier in machine learning, see maximum | entropy classifier. |
Maximum | entropy classifiers are commonly used as alternatives |
t likely state of the universe as one of high | entropy, closer to uniform and without order. |
The zigzag | entropy coding used in JPEG images is evocative of th |
The article on Shannon's information | entropy contains a good explanation of the discrete c |
Penrose associates the initial low | entropy content of the Universe with the effective va |
these problems is rooted in a concept of the | entropy content of gravitational fields. |
singularity (the Big Bang), he proposes, the | entropy content of the cosmological gravitational fie |
If | entropy continues to increase in the contracting phas |
d solubility is determined by the unfavorable | entropy contribution due to the ordering of the water |
∆Smc is the | entropy cost for fixing the backbone in the folded st |
e particle into a computing apparatus with no | entropy cost; but only if the apparatus has already b |
ameter appears in the denominator because the | entropy counts the number of edges puncturing the eve |
ressions and hence determine the enthalpy and | entropy departures. |
The concept of | entropy developed in response to the observation that |
A temperature | entropy diagram, or T-s diagram, is used in thermodyn |
This definition of | entropy does not have a clear thermodynamical interpr |
ts are quantized and the quantized values are | entropy encoded and/or run length encoded. |
ree energy perturbation (FEP), quasi-harmonic | entropy estimation, correlation analysis and combined |
t the New Wave, based on his Ph.D thesis, The | Entropy Exhibition: Michael Moorcock and the British |
The | Entropy Exhibition: Michael Moorcock and the British |
ead confusion and can hinder comprehension of | entropy for most students. |
The information gain is equal to the total | entropy for an attribute if for each of the attribute |
It is also called the relative | entropy, for using Q instead of P. |
The Maximum | Entropy Formalism (ed. |
It can be shown that the Gibbs | entropy formula, with the natural logarithm, reproduc |
ents (CVEs) as a means of excising off excess | entropy from this universe into others as to stave of |
natural extension of the principle of maximum | entropy from discrete to continuous distributions, fo |
H(p) is distinguished from the | entropy function by its taking a single scalar consta |
It also features an | entropy gathering utility, coming in different versio |
xchange design based on two original methods: | entropy generation minimization, and constructal theo |
ructal law of design and evolution in nature, | entropy generation minimization, scale analysis of co |
is the | entropy generation contributed by heat transfer |
is the | entropy generation contributed by fluid friction. |
Volume | entropy h is always bounded above by the topological |
The BGS | entropy has been used in the context of quantum gravi |
ation, the system should exhibit no change in | entropy, i.e. |
d) in the UK, and on Russo's own South Cherry | Entropy imprint of Virtual Label digitally in North A |
fficient when sorting data which exhibits low | entropy, in effect where the data is well ordered or |
ible for an overall increase in the amount of | entropy in the Universe, and so inducing a cosmologic |
S - specific | entropy in kJ per kilogram-kelvin |
second law of thermodynamics states that the | entropy in the universe will always increase. |
It is defined as b raised to the power of | entropy in base b, or more often as b raised to the p |
his critique of the inadequacy of describing | entropy in terms of “disorder”. |
resources for chemistry instructors, equating | entropy increase as the spontaneous dispersal of ener |
will be no net exchange of heat or work - the | entropy increase will be entirely due to the mixing o |
ures, there will, of course, be an additional | entropy increase due to these differences being equil |
A special case of | entropy increase, the entropy of mixing, occurs when |
irections of the atomic magnets represents an | entropy increase? |
esks, and Disorderly Dorm Rooms - Examples of | Entropy Increase? |
ut never reaches thermodynamic equilibrium as | entropy increases continuously without limit due to t |
Katok's | entropy inequality was recently exploited to obtain a |
Furthermore, if the | entropy integral on the right-hand side converges, th |
The standard molar | entropy is usually given the symbol S°, and the units |
Therefore configuration | entropy is the same as macroscopic entropy. |
Its differential | entropy is then |
E.g. its | entropy is captured by a law directly analogous to th |
The configurational | entropy is also known as microscopic entropy or confo |
The relative | entropy is not a metric. |
The Tsallis | entropy is defined as |
This residual | entropy is often quite negligible. |
In quantum information theory, the joint | entropy is generalized into the joint quantum entropy |
The configurational | entropy is related to the number of possible configur |
In general, configurational | entropy is the foundation of statistical thermodynami |
The term Boltzmann | entropy is also sometimes used to indicate entropies |
The von Neumann | entropy is also strongly subadditive. |
Loop | entropy is the entropy lost upon bringing together tw |
Conformational | entropy is the entropy associated with the physical a |
In chemistry, the standard molar | entropy is the entropy content of one mole of substan |
Residual | entropy is small amount of entropy which is present e |
al, term k is zero, and therefore the Tsallis | entropy is in closed-form. |
Joint | entropy is a measure of the uncertainty associated wi |
The principle of maximum | entropy is useful explicitly only when applied to tes |
In statistical mechanics, configuration | entropy is the portion of a system's entropy that is |
When the chain is stretched, the | entropy is reduced by a large margin because there ar |
, because Q is not a state function while the | entropy is. |
y is always less than or equal to the Shannon | entropy; it is equal when all the probabilities pi ar |
ot an expanding universe can approach maximal | entropy; it has been proposed that in an expanding un |
rona later resurfaces as the manifestation of | Entropy itself and battles the New Guardians. |
tropic or particle isentropic flow, where the | entropy level of each fluid particle does not change |
As an example, consider the | entropy lost upon making the contacts between residue |
In machine learning, a maximum | entropy Markov model (MEMM), or conditional Markov mo |
The loop | entropy may also vary with the position of the contac |
The coherent information is an | entropy measure used in quantum information theory. |
perature to the increase in the dimensionless | entropy measured in bits. |
The maximum | entropy method applied to spectral density estimation |
vation of the partition function with maximum | entropy methods. |
Other programs include: | Entropy, Mixmaster, GPG, NEWSPOST, plus Plugins for m |
ting the known information allows the maximum | entropy model to derive a better estimate of the cent |
It is simply the application of maximum | entropy modeling to any type of spectrum and is used |
In maximum | entropy modeling probability distributions are create |
ches a low-density state, it recontracts, but | entropy now decreases, pointing the thermodynamic arr |
For ε > 0, denote by N(T, dX; ε) the | entropy number, i.e. the minimal number of (open) dX- |
tes corresponding to a horizon of area A, the | entropy of black holes is seen to be equal |
fficient Reason, and the Principle of Maximum | Entropy of E.T. Jaynes. |
The joint | entropy of a set of variables is greater than or equa |
The joint | entropy of a set of variables is less than or equal t |
index of 1 indicates that the distributional | entropy of the system under investigation is almost s |
The | entropy of a pure crystalline structure can be 0 J mo |
gation of hydrocarbons, negative enthalpy and | entropy of dissolution of gases, etc. |
density function f : Rn → R, the information | entropy of X, denoted h(X), is defined to be |
means that the fractional perturbation in the | entropy of each species of particle is equal. |
Thus in situations such as in the | entropy of mixing when the two or more different subs |
third law of thermodynamics which states, the | entropy of a system at absolute zero is a well-define |
kur-Tetrode equation is an expression for the | entropy of a monatomic classical ideal gas which inco |
the Tsallis | entropy of this system satisfies |
Estimating the differential | entropy of a system or process, given some observatio |
I(X;A) of X and A - i.e. the reduction in the | entropy of X achieved by learning the state of the ra |
The | entropy of fusion is the increase in entropy when mel |
The | entropy of vaporization is the increase in entropy wh |
The relative | entropy of entanglement of ρ is defined by |
with Strominger, that the Bekenstein-Hawking | entropy of a black hole can be accounted for by solit |
The topological | entropy of the map f is defined by |
Adding the | entropy of transition from α-AgI to β-AgI to the entr |
Put in words, the information | entropy of a distribution P is less than or equal to |
n theorized as being responsible for the high | entropy of black holes; while the likelihood of any g |
The measure of this disorder is called the | entropy of the system. |
The information | entropy of a random event is the expected value of it |
Sinai demonstrated that the Kolmogorov | entropy of a Bernoulli scheme is given by |
riments, it is quite difficult to measure the | entropy of a system. |
The | entropy of the interior region is bounded by the surf |
quantum statistical mechanical origin of the | entropy of a black hole and the entropy of a cosmolog |
ster conformal field theory, and computed the | entropy of BTZ black holes. |
where H(X) is | entropy of X and I(X,Y) is mutual information between |
s article for the constraints placed upon the | entropy of an ideal gas by thermodynamics alone. |
It is driven by the | entropy of the overall reaction, as the byproducts Me |
The | entropy of fusion (melting) for α-AgI is approximatel |
s approximation, is to start with the Shannon | entropy or compositional uncertainty |
mical invariants as correlation dimension, K2 | entropy or mutual information, which are independent |
they may be able to "flush" all the existing | entropy out of the system and put it into a known sta |
emperature and approaching a state of maximal | entropy over a very long time period. |
Moving mirrors create | entropy, particles, energy and gravitational-like eff |
In mathematics, the | entropy power inequality is a result in probability t |
It shows that the | entropy power of suitably well-behaved random variabl |
and the | entropy power of X, denoted N(X), is defined to be |
こんにちは ゲスト さん
ログイン |
Weblio会員(無料)になると
|
こんにちは ゲスト さん
ログイン |
Weblio会員(無料)になると
|