Topic: Computing/Computer science (Page 3)

You are looking at all articles with the topic "Computing/Computer science". We found 35 matches.

Hint: To view all topics, click here. Too see the most popular topics, click here instead.

πŸ”— The SOLID principles of object-oriented design

πŸ”— Computing πŸ”— Computing/Software πŸ”— Computing/Computer science

In object-oriented computer programming, SOLID is a mnemonic acronym for five design principles intended to make software designs more understandable, flexible and maintainable. It is not related to the GRASP software design principles. The principles are a subset of many principles promoted by American software engineer and instructor Robert C. Martin. Though they apply to any object-oriented design, the SOLID principles can also form a core philosophy for methodologies such as agile development or adaptive software development. The theory of SOLID principles was introduced by Martin in his 2000 paper Design Principles and Design Patterns, although the SOLID acronym was introduced later by Michael Feathers.

Discussed on

πŸ”— Occam programming language

πŸ”— Computing πŸ”— Computing/Software πŸ”— Computing/Computer science

occam is a programming language which is concurrent and builds on the communicating sequential processes (CSP) process algebra, and shares many of its features. It is named after philosopher William of Ockham after whom Occam's razor is named.

occam is an imperative procedural language (such as Pascal). It was developed by David May and others at Inmos (trademark INMOS), advised by Tony Hoare, as the native programming language for their transputer microprocessors, but implementations for other platforms are available. The most widely known version is occam 2; its programming manual was written by Steven Ericsson-Zenith and others at Inmos.

Discussed on

πŸ”— Merkle Tree

πŸ”— Computing πŸ”— Computing/Software πŸ”— Computing/Computer science πŸ”— Cryptography πŸ”— Cryptography/Computer science

In cryptography and computer science, a hash tree or Merkle tree is a tree in which every leaf node is labelled with the cryptographic hash of a data block, and every non-leaf node is labelled with the cryptographic hash of the labels of its child nodes. Hash trees allow efficient and secure verification of the contents of large data structures. Hash trees are a generalization of hash lists and hash chains.

Demonstrating that a leaf node is a part of a given binary hash tree requires computing a number of hashes proportional to the logarithm of the number of leaf nodes of the tree; this contrasts with hash lists, where the number is proportional to the number of leaf nodes itself.

The concept of hash trees is named after Ralph Merkle, who patented it in 1979.

Discussed on

πŸ”— Mark V. Shaney

πŸ”— Computing πŸ”— Computing/Software πŸ”— Linguistics πŸ”— Computing/Computer science πŸ”— Linguistics/Applied Linguistics

Mark V. Shaney is a synthetic Usenet user whose postings in the net.singles newsgroups were generated by Markov chain techniques, based on text from other postings. The username is a play on the words "Markov chain". Many readers were fooled into thinking that the quirky, sometimes uncannily topical posts were written by a real person.

The system was designed by Rob Pike with coding by Bruce Ellis. Don P. Mitchell wrote the Markov chain code, initially demonstrating it to Pike and Ellis using the Tao Te Ching as a basis. They chose to apply it to the net.singles netnews group.

Discussed on

πŸ”— Comb Sort - Just As Good As Quick Sort

πŸ”— Computing πŸ”— Computer science πŸ”— Computing/Software πŸ”— Computing/Computer science

Comb sort is a relatively simple sorting algorithm originally designed by WΕ‚odzimierz Dobosiewicz and Artur Borowy in 1980, later rediscovered by Stephen Lacey and Richard Box in 1991. Comb sort improves on bubble sort.

Discussed on

πŸ”— Non-blocking synchronization

πŸ”— Computing πŸ”— Computing/Computer science

In computer science, an algorithm is called non-blocking if failure or suspension of any thread cannot cause failure or suspension of another thread; for some operations, these algorithms provide a useful alternative to traditional blocking implementations. A non-blocking algorithm is lock-free if there is guaranteed system-wide progress, and wait-free if there is also guaranteed per-thread progress.

The word "non-blocking" was traditionally used to describe telecommunications networks that could route a connection through a set of relays "without having to re-arrange existing calls", see Clos network. Also, if the telephone exchange "is not defective, it can always make the connection", see nonblocking minimal spanning switch.

Discussed on

πŸ”— The Complexity of Songs

πŸ”— Computing πŸ”— Computing/Software πŸ”— Computing/Computer science

"The Complexity of Songs" is a journal article published by computer scientist Donald Knuth in 1977, as an in-joke about computational complexity theory. The article capitalizes on the tendency of popular songs to devolve from long and content-rich ballads to highly repetitive texts with little or no meaningful content. The article notes that a song of length N words may be produced remembering, e.g., only O(log N) words ("space complexity" of the song).

Discussed on

πŸ”— John von Neumann

πŸ”— Biography πŸ”— Computing πŸ”— Mathematics πŸ”— Military history πŸ”— Military history/North American military history πŸ”— Military history/United States military history πŸ”— Military history/Military science, technology, and theory πŸ”— Physics πŸ”— Economics πŸ”— Philosophy πŸ”— Philosophy/Logic πŸ”— Biography/science and academia πŸ”— Philosophy/Philosophy of science πŸ”— Philosophy/Contemporary philosophy πŸ”— Military history/Military biography πŸ”— Biography/military biography πŸ”— History of Science πŸ”— Computing/Computer science πŸ”— Philosophy/Philosophers πŸ”— Education πŸ”— Hungary πŸ”— Military history/World War II πŸ”— Military history/Cold War πŸ”— Physics/History πŸ”— Physics/Biographies πŸ”— Game theory πŸ”— Eastern Europe

John von Neumann (; Hungarian: Neumann JΓ‘nos Lajos, pronouncedΒ [ˈnΙ’jmΙ’n ˈjaːnoΚƒ ˈlΙ’joΚƒ]; December 28, 1903 – FebruaryΒ 8, 1957) was a Hungarian-American mathematician, physicist, computer scientist, engineer and polymath. Von Neumann was generally regarded as the foremost mathematician of his time and said to be "the last representative of the great mathematicians"; who integrated both pure and applied sciences.

He made major contributions to a number of fields, including mathematics (foundations of mathematics, functional analysis, ergodic theory, representation theory, operator algebras, geometry, topology, and numerical analysis), physics (quantum mechanics, hydrodynamics, and quantum statistical mechanics), economics (game theory), computing (Von Neumann architecture, linear programming, self-replicating machines, stochastic computing), and statistics.

He was a pioneer of the application of operator theory to quantum mechanics in the development of functional analysis, and a key figure in the development of game theory and the concepts of cellular automata, the universal constructor and the digital computer.

He published over 150 papers in his life: about 60 in pure mathematics, 60 in applied mathematics, 20 in physics, and the remainder on special mathematical subjects or non-mathematical ones. His last work, an unfinished manuscript written while he was in hospital, was later published in book form as The Computer and the Brain.

His analysis of the structure of self-replication preceded the discovery of the structure of DNA. In a short list of facts about his life he submitted to the National Academy of Sciences, he stated, "The part of my work I consider most essential is that on quantum mechanics, which developed in GΓΆttingen in 1926, and subsequently in Berlin in 1927–1929. Also, my work on various forms of operator theory, Berlin 1930 and Princeton 1935–1939; on the ergodic theorem, Princeton, 1931–1932."

During World War II, von Neumann worked on the Manhattan Project with theoretical physicist Edward Teller, mathematician StanisΕ‚aw Ulam and others, problem solving key steps in the nuclear physics involved in thermonuclear reactions and the hydrogen bomb. He developed the mathematical models behind the explosive lenses used in the implosion-type nuclear weapon, and coined the term "kiloton" (of TNT), as a measure of the explosive force generated.

After the war, he served on the General Advisory Committee of the United States Atomic Energy Commission, and consulted for a number of organizations, including the United States Air Force, the Army's Ballistic Research Laboratory, the Armed Forces Special Weapons Project, and the Lawrence Livermore National Laboratory. As a Hungarian Γ©migrΓ©, concerned that the Soviets would achieve nuclear superiority, he designed and promoted the policy of mutually assured destruction to limit the arms race.

Discussed on

πŸ”— Unix philosophy

πŸ”— Computing πŸ”— Computing/Software πŸ”— Computing/Computer science πŸ”— Linux

The Unix philosophy, originated by Ken Thompson, is a set of cultural norms and philosophical approaches to minimalist, modular software development. It is based on the experience of leading developers of the Unix operating system. Early Unix developers were important in bringing the concepts of modularity and reusability into software engineering practice, spawning a "software tools" movement. Over time, the leading developers of Unix (and programs that ran on it) established a set of cultural norms for developing software; these norms became as important and influential as the technology of Unix itself; this has been termed the "Unix philosophy."

The Unix philosophy emphasizes building simple, short, clear, modular, and extensible code that can be easily maintained and repurposed by developers other than its creators. The Unix philosophy favors composability as opposed to monolithic design.

Discussed on

πŸ”— Bitcoin Cryptocurrency

πŸ”— Internet πŸ”— Computing πŸ”— Computing/Computer hardware πŸ”— Finance & Investment πŸ”— Economics πŸ”— Law πŸ”— Computing/Software πŸ”— Computing/Free and open-source software πŸ”— Computing/Computer science πŸ”— Cryptography πŸ”— Cryptography/Computer science πŸ”— Numismatics πŸ”— Guild of Copy Editors πŸ”— Numismatics/Cryptocurrency πŸ”— Cryptocurrency πŸ”— Open πŸ”— Computing/Computer Security

Bitcoin (β‚Ώ) is a cryptocurrency. It is a decentralized digital currency without a central bank or single administrator that can be sent from user to user on the peer-to-peer bitcoin network without the need for intermediaries.

Transactions are verified by network nodes through cryptography and recorded in a public distributed ledger called a blockchain. Bitcoin was invented in 2008 by an unknown person or group of people using the name Satoshi Nakamoto and started in 2009 when its source code was released as open-source software. Bitcoins are created as a reward for a process known as mining. They can be exchanged for other currencies, products, and services. Research produced by University of Cambridge estimates that in 2017, there were 2.9 to 5.8 million unique users using a cryptocurrency wallet, most of them using bitcoin.

Bitcoin has been criticized for its use in illegal transactions, its high electricity consumption, price volatility, and thefts from exchanges. Some economists, including several Nobel laureates, have characterized it as a speculative bubble. Bitcoin has also been used as an investment, although several regulatory agencies have issued investor alerts about bitcoin.