New Articles (Page 276)

To stay up to date you can also follow on Mastodon.

πŸ”— Hapax legomenon

πŸ”— Religion πŸ”— Bible πŸ”— Linguistics

In corpus linguistics, a hapax legomenon ( also or ; pl. hapax legomena; sometimes abbreviated to hapax) is a word that occurs only once within a context, either in the written record of an entire language, in the works of an author, or in a single text. The term is sometimes incorrectly used to describe a word that occurs in just one of an author's works, but more than once in that particular work. Hapax legomenon is a transliteration of Greek ἅπαξ λΡγόμΡνον, meaning "(something) being said (only) once".

The related terms dis legomenon, tris legomenon, and tetrakis legomenon respectively (, , ) refer to double, triple, or quadruple occurrences, but are far less commonly used.

Hapax legomena are quite common, as predicted by Zipf's law, which states that the frequency of any word in a corpus is inversely proportional to its rank in the frequency table. For large corpora, about 40% to 60% of the words are hapax legomena, and another 10% to 15% are dis legomena. Thus, in the Brown Corpus of American English, about half of the 50,000 distinct words are hapax legomena within that corpus.

Hapax legomenon refers to a word's appearance in a body of text, not to either its origin or its prevalence in speech. It thus differs from a nonce word, which may never be recorded, may find currency and may be widely recorded, or may appear several times in the work which coins it, and so on.

Discussed on

πŸ”— Wallpaper group

πŸ”— Mathematics

A wallpaper group (or plane symmetry group or plane crystallographic group) is a mathematical classification of a two-dimensional repetitive pattern, based on the symmetries in the pattern. Such patterns occur frequently in architecture and decorative art, especially in textiles and tiles as well as wallpaper.

The simplest wallpaper group, Group p1, applies when there is no symmetry other than the fact that a pattern repeats over regular intervals in two dimensions, as shown in the section on p1 below.

Consider the following examples of patterns with more forms of symmetry:

Examples A and B have the same wallpaper group; it is called p4m in the IUC notation and *442 in the orbifold notation. Example C has a different wallpaper group, called p4g or 4*2 . The fact that A and B have the same wallpaper group means that they have the same symmetries, regardless of details of the designs, whereas C has a different set of symmetries despite any superficial similarities.

The number of symmetry groups depends on the number of dimensions in the patterns. Wallpaper groups apply to the two-dimensional case, intermediate in complexity between the simpler frieze groups and the three-dimensional space groups. Subtle differences may place similar patterns in different groups, while patterns that are very different in style, color, scale or orientation may belong to the same group.

A proof that there were only 17 distinct groups of such planar symmeries was first carried out by Evgraf Fedorov in 1891 and then derived independently by George PΓ³lya in 1924. The proof that the list of wallpaper groups was complete only came after the much harder case of space groups had been done. The seventeen possible wallpaper groups are listed below in Β§Β The seventeen groups.

Discussed on

πŸ”— Ship's cat

πŸ”— Cats πŸ”— Ships πŸ”— Rodents

The ship's cat has been a common feature on many trading, exploration, and naval ships dating to ancient times. Cats have been carried on ships for many reasons, most importantly to control rodents. Vermin aboard a ship can cause damage to ropes, woodwork, and more recently, electrical wiring. Also, rodents threaten ships' stores, devour crews' foodstuff, and could cause economic damage to ships' cargo such as grain. They are also a source of disease, which is dangerous for ships that are at sea for long periods of time. Rat fleas are carriers of plague, and rats on ships were believed to be a primary vector of the Black Death.

Cats naturally attack and kill rodents, and their natural ability to adapt to new surroundings made them suitable for service on a ship. In addition, they offer companionship and a sense of home, security and camaraderie to sailors away from home.

Discussed on

πŸ”— Secretary Problem

πŸ”— Mathematics πŸ”— Statistics

The secretary problem is a problem that demonstrates a scenario involving optimal stopping theory. The problem has been studied extensively in the fields of applied probability, statistics, and decision theory. It is also known as the marriage problem, the sultan's dowry problem, the fussy suitor problem, the googol game, and the best choice problem.

The basic form of the problem is the following: imagine an administrator who wants to hire the best secretary out of n {\displaystyle n} rankable applicants for a position. The applicants are interviewed one by one in random order. A decision about each particular applicant is to be made immediately after the interview. Once rejected, an applicant cannot be recalled. During the interview, the administrator gains information sufficient to rank the applicant among all applicants interviewed so far, but is unaware of the quality of yet unseen applicants. The question is about the optimal strategy (stopping rule) to maximize the probability of selecting the best applicant. If the decision can be deferred to the end, this can be solved by the simple maximum selection algorithm of tracking the running maximum (and who achieved it), and selecting the overall maximum at the end. The difficulty is that the decision must be made immediately.

The shortest rigorous proof known so far is provided by the odds algorithm (Bruss 2000). It implies that the optimal win probability is always at least 1 / e {\displaystyle 1/e} (where e is the base of the natural logarithm), and that the latter holds even in a much greater generality (2003). The optimal stopping rule prescribes always rejecting the first ∼ n / e {\displaystyle \sim n/e} applicants that are interviewed and then stopping at the first applicant who is better than every applicant interviewed so far (or continuing to the last applicant if this never occurs). Sometimes this strategy is called the 1 / e {\displaystyle 1/e} stopping rule, because the probability of stopping at the best applicant with this strategy is about 1 / e {\displaystyle 1/e} already for moderate values of n {\displaystyle n} . One reason why the secretary problem has received so much attention is that the optimal policy for the problem (the stopping rule) is simple and selects the single best candidate about 37% of the time, irrespective of whether there are 100 or 100 million applicants.

Discussed on

πŸ”— Wikipedia: Are you evil?

Congratulations, you have discovered Wikipedia, one of the most popular sites on the Internet and probably the leading informational resource in the world today.

Unfortunately some people are evil and have to be banned. In order to save time, we have compiled this easy questionnaire. It's multiple choice, one answer only.

Q: You have discovered a website which has enormous reach and is used by millions of people every day. You think:

  1. Wow, this is awesome! How can I help?
  2. Wow, this is awesome! How can I use this to my advantage?

Scoring:

  1. 1 point
  2. 0 points

Marking: 0 points: Fail. Please die in a fire. 1 point: Pass. Welcome!


It seems like the authors of this 'test' are evil, willing people to die and in a most horrible manner.

Discussed on

πŸ”— Therac-25

πŸ”— Medicine

The Therac-25 was a computer-controlled radiation therapy machine produced by Atomic Energy of Canada Limited (AECL) in 1982 after the Therac-6 and Therac-20 units (the earlier units had been produced in partnership with CGR of France).

It was involved in at least six accidents between 1985 and 1987, in which patients were given massive overdoses of radiation. Because of concurrent programming errors, it sometimes gave its patients radiation doses that were hundreds of times greater than normal, resulting in death or serious injury. These accidents highlighted the dangers of software control of safety-critical systems, and they have become a standard case study in health informatics and software engineering. Additionally the overconfidence of the engineers and lack of proper due diligence to resolve reported software bugs are highlighted as an extreme case where the engineers' overconfidence in their initial work and failure to believe the end users' claims caused drastic repercussions.

Discussed on

πŸ”— The fastest pulsar spins at 716Hz; its equator spins at 24% the speed of light

πŸ”— Astronomy πŸ”— Astronomy/Astronomical objects

PSR J1748βˆ’2446ad is the fastest-spinning pulsar known, at 716 Hz, or 716 times per second. This pulsar was discovered by Jason W. T. Hessels of McGill University on November 10, 2004 and confirmed on January 8, 2005.

If the neutron star is assumed to contain less than two times the mass of the Sun, within the typical range of neutron stars, its radius is constrained to be less than 16Β km. At its equator it is spinning at approximately 24% of the speed of light, or over 70,000Β km per second.

The pulsar is located in a globular cluster of stars called Terzan 5, located approximately 18,000 light-years from Earth in the constellation Sagittarius. It is part of a binary system and undergoes regular eclipses with an eclipse magnitude of about 40%. Its orbit is highly circular with a 26-hour period. The other object is at least 0.14 solar masses, with a radius of 5–6 solar radii. Hessels et al. state that the companion may be a "bloated main-sequence star, possibly still filling its Roche Lobe". Hessels et al. go on to speculate that gravitational radiation from the pulsar might be detectable by LIGO.

πŸ”— Chinese restaurant process

πŸ”— Statistics

In probability theory, the Chinese restaurant process is a discrete-time stochastic process, analogous to seating customers at tables in a Chinese restaurant. Imagine a Chinese restaurant with an infinite number of circular tables, each with infinite capacity. Customer 1 sits at the first table. The next customer either sits at the same table as customer 1, or the next table. This continues, with each customer choosing to either sit at an occupied table with a probability proportional to the number of customers already there (i.e., they are more likely to sit at a table with many customers than few), or an unoccupied table. At time n, the n customers have been partitioned among m ≀ n tables (or blocks of the partition). The results of this process are exchangeable, meaning the order in which the customers sit does not affect the probability of the final distribution. This property greatly simplifies a number of problems in population genetics, linguistic analysis, and image recognition.

David J. Aldous attributes the restaurant analogy to Jim Pitman and Lester Dubins in his 1983 book.

Discussed on

πŸ”— Kernel Embedding of Distributions

In machine learning, the kernel embedding of distributions (also called the kernel mean or mean map) comprises a class of nonparametric methods in which a probability distribution is represented as an element of a reproducing kernel Hilbert space (RKHS). A generalization of the individual data-point feature mapping done in classical kernel methods, the embedding of distributions into infinite-dimensional feature spaces can preserve all of the statistical features of arbitrary distributions, while allowing one to compare and manipulate distributions using Hilbert space operations such as inner products, distances, projections, linear transformations, and spectral analysis. This learning framework is very general and can be applied to distributions over any space Ξ© {\displaystyle \Omega } on which a sensible kernel function (measuring similarity between elements of Ξ© {\displaystyle \Omega } ) may be defined. For example, various kernels have been proposed for learning from data which are: vectors in R d {\displaystyle \mathbb {R} ^{d}} , discrete classes/categories, strings, graphs/networks, images, time series, manifolds, dynamical systems, and other structured objects. The theory behind kernel embeddings of distributions has been primarily developed by Alex Smola, Le Song , Arthur Gretton, and Bernhard SchΓΆlkopf. A review of recent works on kernel embedding of distributions can be found in.

The analysis of distributions is fundamental in machine learning and statistics, and many algorithms in these fields rely on information theoretic approaches such as entropy, mutual information, or Kullback–Leibler divergence. However, to estimate these quantities, one must first either perform density estimation, or employ sophisticated space-partitioning/bias-correction strategies which are typically infeasible for high-dimensional data. Commonly, methods for modeling complex distributions rely on parametric assumptions that may be unfounded or computationally challenging (e.g. Gaussian mixture models), while nonparametric methods like kernel density estimation (Note: the smoothing kernels in this context have a different interpretation than the kernels discussed here) or characteristic function representation (via the Fourier transform of the distribution) break down in high-dimensional settings.

Methods based on the kernel embedding of distributions sidestep these problems and also possess the following advantages:

  1. Data may be modeled without restrictive assumptions about the form of the distributions and relationships between variables
  2. Intermediate density estimation is not needed
  3. Practitioners may specify the properties of a distribution most relevant for their problem (incorporating prior knowledge via choice of the kernel)
  4. If a characteristic kernel is used, then the embedding can uniquely preserve all information about a distribution, while thanks to the kernel trick, computations on the potentially infinite-dimensional RKHS can be implemented in practice as simple Gram matrix operations
  5. Dimensionality-independent rates of convergence for the empirical kernel mean (estimated using samples from the distribution) to the kernel embedding of the true underlying distribution can be proven.
  6. Learning algorithms based on this framework exhibit good generalization ability and finite sample convergence, while often being simpler and more effective than information theoretic methods

Thus, learning via the kernel embedding of distributions offers a principled drop-in replacement for information theoretic approaches and is a framework which not only subsumes many popular methods in machine learning and statistics as special cases, but also can lead to entirely new learning algorithms.

Discussed on

πŸ”— Edward Bernays

πŸ”— United States πŸ”— Biography πŸ”— Marketing & Advertising πŸ”— Judaism πŸ”— Jewish history πŸ”— Vienna

Edward Louis Bernays (; German: [bΙ›ΙΜ―ΛˆnaΙͺs]; November 22, 1891 βˆ’ March 9, 1995) was an Austrian-American pioneer in the field of public relations and propaganda, referred to in his obituary as "the father of public relations". Bernays was named one of the 100 most influential Americans of the 20th century by Life. He was the subject of a full length biography by Larry Tye called The Father of Spin (1999) and later an award-winning 2002 documentary for the BBC by Adam Curtis called The Century of the Self.

His best-known campaigns include a 1929 effort to promote female smoking by branding cigarettes as feminist "Torches of Freedom" and his work for the United Fruit Company connected with the CIA-orchestrated overthrow of the democratically elected Guatemalan government in 1954. He worked for dozens of major American corporations including Procter & Gamble and General Electric, and for government agencies, politicians, and non-profit organizations.

Of his many books, Crystallizing Public Opinion (1923) and Propaganda (1928) gained special attention as early efforts to define and theorize the field of public relations. Citing works of writers such as Gustave Le Bon, Wilfred Trotter, Walter Lippmann, and his own double uncle Sigmund Freud, he described the masses as irrational and subject to herd instinctβ€”and outlined how skilled practitioners could use crowd psychology and psychoanalysis to control them in desirable ways.

Discussed on