Computational science in the context of Computational physics


Computational science in the context of Computational physics

Computational science Study page number 1 of 1

Play TriviaQuestions Online!

or

Skip to study material about Computational science in the context of "Computational physics"


⭐ Core Definition: Computational science

Computational science, also known as scientific computing, technical computing or scientific computation (SC), is a division of science, and more specifically the computer sciences, which uses advanced computing capabilities to understand and solve complex physical problems in science. While this typically extends into computational specializations, this field of study includes:

In practical use, it is typically the application of computer simulation and other forms of computation from numerical analysis and theoretical computer science to solve problems in various scientific disciplines. The field is different from theory and laboratory experiments, which are the traditional forms of science and engineering. The scientific computing approach is to gain understanding through the analysis of mathematical models implemented on computers. Scientists and engineers develop computer programs and application software that model systems being studied and run these programs with various sets of input parameters. The essence of computational science is the application of numerical algorithms and computational mathematics. In some cases, these models require massive amounts of calculations (usually floating-point) and are often executed on supercomputers or distributed computing platforms.

↓ Menu
HINT:

πŸ‘‰ Computational science in the context of Computational physics

Computational physics is the study and implementation of numerical analysis to solve problems in physics. Historically, computational physics was the first application of modern computers in science, and is now a subset of computational science. It is sometimes regarded as a subdiscipline (or offshoot) of theoretical physics, but others consider it an intermediate branch between theoretical and experimental physics β€” an area of study which supplements both theory and experiment.

↓ Explore More Topics
In this Dossier

Computational science in the context of Geographic information science

Geographic information science (GIScience, GISc) or geoinformation science is a scientific discipline at the crossroads of computational science, social science, and natural science that studies geographic information, including how it represents phenomena in the real world, how it represents the way humans understand the world, and how it can be captured, organized, and analyzed. It is a sub-field of geography, specifically part of technical geography. It has applications to both physical geography and human geography, although its techniques can be applied to many other fields of study as well as many different industries.

As a field of study or profession, it can be contrasted with geographic information systems (GIS), which are the actual repositories of geospatial data, the software tools for carrying out relevant tasks, and the profession of GIS users. That said, one of the major goals of GIScience is to find practical ways to improve GIS data, software, and professional practice; it is more focused on how GIS is applied in real life as opposed to being a geographic information system tool in and of itself. The field is also sometimes called geographical information science.

View the full Wikipedia page for Geographic information science
↑ Return to Menu

Computational science in the context of Mathematical science

The Mathematical Sciences are a group of areas of study that includes, in addition to mathematics, those academic disciplines that are primarily mathematical in nature but may not be universally considered subfields of mathematics proper.

Statistics, for example, is mathematical in its methods but grew out of bureaucratic and scientific observations, which merged with inverse probability and then grew through applications in some areas of physics, biometrics, and the social sciences to become its own separate, though closely allied, field. Theoretical astronomy, theoretical physics, theoretical and applied mechanics, continuum mechanics, mathematical chemistry, actuarial science, computer science, computational science, data science, operations research, quantitative biology, control theory, econometrics, geophysics and mathematical geosciences are likewise other fields often considered part of the mathematical sciences.

View the full Wikipedia page for Mathematical science
↑ Return to Menu

Computational science in the context of Computational statistics

Computational statistics, or statistical computing, is the study which is the intersection of statistics and computer science, and refers to the statistical methods that are enabled by using computational methods. It is the area of computational science (or scientific computing) specific to the mathematical science of statistics. This area is fast developing. The view that the broader concept of computing must be taught as part of general statistical education is gaining momentum.

As in traditional statistics the goal is to transform raw data into knowledge, but the focus lies on computer intensive statistical methods, such as cases with very large sample size and non-homogeneous data sets.

View the full Wikipedia page for Computational statistics
↑ Return to Menu

Computational science in the context of Workstation

A workstation is a special computer designed for technical or scientific applications. Intended primarily to be used by a single user, they are commonly connected to a local area network and run multi-user operating systems. The term workstation has been used loosely to refer to everything from a mainframe computer terminal to a PC connected to a network, but the most common form refers to the class of hardware offered by several current and defunct companies such as Sun Microsystems, Silicon Graphics, Apollo Computer, DEC, HP, NeXT, and IBM which powered the 3D computer graphics revolution of the late 1990s.

Workstations formerly offered higher performance specifications than mainstream personal computers, especially in terms of processing, graphics, memory, and multitasking. Workstations are optimized for the visualization and manipulation of different types of complex data such as 3D mechanical design, engineering simulations like computational fluid dynamics, animation, video editing, image editing, medical imaging, image rendering, computational science, generating mathematical plots, and software development. Typically, the form factor is that of a desktop computer, which consists of a high-resolution display, a keyboard, and a mouse at a minimum, but also offers multiple displays, graphics tablets, and 3D mice for manipulating objects and navigating scenes. Workstations were the first segment of the computer market to present advanced accessories, and collaboration tools like videoconferencing.

View the full Wikipedia page for Workstation
↑ Return to Menu

Computational science in the context of Machine perception

Machine perception is the capability of a computer system to interpret data in a manner that is similar to the way humans use their senses to relate to the world around them. The basic method that the computers take in and respond to their environment is through the attached hardware. Until recently input was limited to a keyboard, or a mouse, but advances in technology, both in hardware and software, have allowed computers to take in sensory input in a way similar to humans.

Machine perception allows the computer to use this sensory input, as well as conventional computational means of gathering information, to gather information with greater accuracy and to present it in a way that is more comfortable for the user. These include computer vision, machine hearing, machine touch, and machine smelling, as artificial scents are, at a chemical compound, molecular, atomic level, indiscernible and identical.

View the full Wikipedia page for Machine perception
↑ Return to Menu

Computational science in the context of Computational anatomy

Computational anatomy is an interdisciplinary field of biology focused on quantitative investigation and modelling of anatomical shapes variability. It involves the development and application of mathematical, statistical and data-analytical methods for modelling and simulation of biological structures.

The field is broadly defined and includes foundations in anatomy, applied mathematics and pure mathematics, machine learning, computational mechanics, computational science, biological imaging, neuroscience, physics, probability, and statistics; it also has strong connections with fluid mechanics and geometric mechanics. Additionally, it complements newer, interdisciplinary fields like bioinformatics and neuroinformatics in the sense that its interpretation uses metadata derived from the original sensor imaging modalities (of which magnetic resonance imaging is one example). It focuses on the anatomical structures being imaged, rather than the medical imaging devices. It is similar in spirit to the history of computational linguistics, a discipline that focuses on the linguistic structures rather than the sensor acting as the transmission and communication media.

View the full Wikipedia page for Computational anatomy
↑ Return to Menu

Computational science in the context of Supercomputer

A supercomputer is a type of computer with a high level of performance as compared to a general-purpose computer. Supercomputers play an important role in the field of computational science, and are used for a wide range of computationally intensive tasks in various fields including quantum mechanics, weather forecasting, climate research, oil and gas exploration, molecular modeling (computing the structures and properties of chemical compounds, biological macromolecules, polymers, and crystals), and physical simulations (such as simulations of aerodynamics, of the early moments of the universe, and of nuclear weapons). They have been essential in the field of cryptanalysis.

The performance of a supercomputer is commonly measured in floating-point operations per second (FLOPS) instead of million instructions per second (MIPS). Since 2022, exascale supercomputers have existed which can perform over 10Β FLOPS. For comparison, a desktop computer has performance in the range of hundreds of gigaFLOPS (10) to tens of teraFLOPS (10). Since November 2017, all of the world's fastest 500 supercomputers run on Linux-based operating systems. Additional research is being conducted in the United States, the European Union, Taiwan, Japan, and China to build faster, more powerful and technologically superior exascale supercomputers.

View the full Wikipedia page for Supercomputer
↑ Return to Menu

Computational science in the context of Information Processing and Management

Information Processing and Management is a bimonthly peer-reviewed academic journal published by Elsevier covering the field of information and computational sciences applied to management. The journal was established in 1963 as Information Storage and Retrieval, obtaining its current name in 1975.

View the full Wikipedia page for Information Processing and Management
↑ Return to Menu

Computational science in the context of Systems programming

Systems programming, or system programming, is the activity of programming computer system software. The primary distinguishing characteristic of systems programming when compared to application programming is that application programming aims to produce software which provides services to the user directly (e.g. word processor), whereas systems programming aims to produce software and software platforms which provide services to other software, are performance constrained, or both (e.g. operating systems, computational science applications, game engines, industrial automation, and software as a service applications).

Systems programming requires a great degree of hardware awareness. Its goal is to achieve efficient use of available resources, either because the software itself is performance-critical or because even small efficiency improvements directly transform into significant savings of time or money.

View the full Wikipedia page for Systems programming
↑ Return to Menu

Computational science in the context of Molecular engineering

Molecular engineering is an emerging field of study concerned with the design and testing of molecular properties, behavior and interactions in order to assemble better materials, systems, and processes for specific functions. This approach, in which observable properties of a macroscopic system are influenced by direct alteration of a molecular structure, falls into the broader category of β€œbottom-up” design. This field is utmost relevant to Cheminformatics, when related to the research in the Computational Sciences.

Molecular engineering is highly interdisciplinary by nature, encompassing aspects of chemical engineering, materials science, bioengineering, electrical engineering, physics, mechanical engineering, and chemistry. There is also considerable overlap with nanotechnology, in that both are concerned with the behavior of materials on the scale of nanometers or smaller. Given the highly fundamental nature of molecular interactions, there are a plethora of potential application areas, limited perhaps only by one's imagination and the laws of physics. However, some of the early successes of molecular engineering have come in the fields of immunotherapy, synthetic biology, and printable electronics (see molecular engineering applications).

View the full Wikipedia page for Molecular engineering
↑ Return to Menu

Computational science in the context of Numerical linear algebra

Numerical linear algebra, sometimes called applied linear algebra, is the study of how matrix operations can be used to create computer algorithms which efficiently and accurately provide approximate answers to questions in continuous mathematics. It is a subfield of numerical analysis, and a type of linear algebra. Computers use floating-point arithmetic and cannot exactly represent irrational data, so when a computer algorithm is applied to a matrix of data, it can sometimes increase the difference between a number stored in the computer and the true number that it is an approximation of. Numerical linear algebra uses properties of vectors and matrices to develop computer algorithms that minimize the error introduced by the computer, and is also concerned with ensuring that the algorithm is as efficient as possible.

Numerical linear algebra aims to solve problems of continuous mathematics using finite precision computers, so its applications to the natural and social sciences are as vast as the applications of continuous mathematics. It is often a fundamental part of engineering and computational science problems, such as image and signal processing, telecommunication, computational finance, materials science simulations, structural biology, data mining, bioinformatics, and fluid dynamics. Matrix methods are particularly used in finite difference methods, finite element methods, and the modeling of differential equations. Noting the broad applications of numerical linear algebra, Lloyd N. Trefethen and David Bau, III argue that it is "as fundamental to the mathematical sciences as calculus and differential equations", even though it is a comparatively small field. Because many properties of matrices and vectors also apply to functions and operators, numerical linear algebra can also be viewed as a type of functional analysis which has a particular emphasis on practical algorithms.

View the full Wikipedia page for Numerical linear algebra
↑ Return to Menu

Computational science in the context of Fortran

Fortran (/ˈfɔːrtrΓ¦n/; formerly FORTRAN) is a third-generation, compiled, imperative programming language designed for numeric computation and scientific computing.

Fortran was originally developed by IBM with a reference manual being released in 1956; however, the first compilers only began to produce accurate code two years later. Fortran computer programs have been written to support scientific and engineering applications, such as numerical weather prediction, finite element analysis, computational fluid dynamics, plasma physics, geophysics, computational physics, crystallography and computational chemistry. It is a popular language for high-performance computing and is used for programs that benchmark and rank the world's fastest supercomputers.

View the full Wikipedia page for Fortran
↑ Return to Menu