Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Series TitleSeries Title
-
Reading LevelReading Level
-
YearFrom:-To:
-
More FiltersMore FiltersContent TypeItem TypeIs Full-Text AvailableSubjectPublisherSourceDonorLanguagePlace of PublicationContributorsLocation
Done
Filters
Reset
58,468
result(s) for
"Mathematical Software"
Sort by:
SciPy 1.0: fundamental algorithms for scientific computing in Python
2020
SciPy is an open-source scientific computing library for the Python programming language. Since its initial release in 2001, SciPy has become a de facto standard for leveraging scientific algorithms in Python, with over 600 unique code contributors, thousands of dependent packages, over 100,000 dependent repositories and millions of downloads per year. In this work, we provide an overview of the capabilities and development practices of SciPy 1.0 and highlight some recent technical developments.
This Perspective describes the development and capabilities of SciPy 1.0, an open source scientific computing library for the Python programming language.
Journal Article
Business models in the software industry : the impact on firm and M&A performance
The relevance of software business models has tremendously increased in recent years. Markus Schief explores opportunities to improve the management of these models. Based on a conceptual framework of software business model characteristics, he conducts large empirical studies to examine the current state of business models in the software industry.
The Accuracy of Computational Results from Wolfram Mathematica in the Context of Summation in Trigonometry
2023
This article explores the accessibility of symbolic computations, such as using the Wolfram Mathematica environment, in promoting the shift from informal experimentation to formal mathematical justifications. We investigate the accuracy of computational results from mathematical software in the context of a certain summation in trigonometry. In particular, the key issue addressed here is the calculated sum ∑n=044tan1+4n°. This paper utilizes Wolfram Mathematica to handle the irrational numbers in the sum more accurately, which it achieves by representing them symbolically rather than using numerical approximations. Can we rely on the calculated result from Wolfram, especially if almost all the addends are irrational, or must the students eventually prove it mathematically? It is clear that the problem can be solved using software; however, the nature of the result raises questions about its correctness, and this inherent informality can encourage a few students to seek viable mathematical proofs. In this way, a balance is reached between formal and informal mathematics.
Journal Article
Code generation with Roslyn
\"Learn how Roslyn's new code generation capability will let you write software that is more concise, runs faster, and is easier to maintain. You will learn from real-world business applications to create better software by letting the computer write its own code based on your business logic already defined in lookup tables. Code Generation with Rosyln is the first book to cover this new capability. You will learn how these techniques can be used to simplify systems integration so that if one system already defines business logic through lookup tables, you can integrate a new system and share business logic by allowing the new system to write its own business logic based on already existing table-based business logic. One of the many benefits you will discover is that Roslyn uses an innovative approach to compiler design, opening up the inner workings of the compiler process. You will learn how to see the syntax tree that Roslyn is building as it compiles your code. Additionally, you will learn to feed it your own syntax tree that you create on the fly.\"-- Provided by publisher
Approximate contact factorization of germs of plane curves
Given an algebraic germ of a plane curve at the origin, in terms of a bivariate polynomial, we analyze the complexity of computing an irreducible decomposition up to any given truncation order. With a suitable representation of the irreducible components, and whenever the characteristic of the ground field is zero or larger than the degree of the germ, we design a new algorithm that involves a nearly linear number of arithmetic operations in the ground field plus a small amount of irreducible univariate polynomial factorizations.
Journal Article
Introduction to computational modeling using C and open-source tools
\"Because computational models usually require high-performance computing to solve large and complex scientific problems, this book presents an introduction to computational models and their implementation using the C programming language. Its primary goal is to present basic to semi-advanced principles of computational modeling, at the level of Calculus and Intermediate Programming. Emphasis is on reasoning about problems, conceptualizing the problem, mathematical modeling, and their computational solution that involves computing results and visualization. \"-- Provided by publisher.
Brian 2, an intuitive and efficient neural simulator
by
Stimberg, Marcel
,
Goodman, Dan FM
,
Brette, Romain
in
Artificial neural networks
,
Brain
,
Cognitive science
2019
Brian 2 allows scientists to simply and efficiently simulate spiking neural network models. These models can feature novel dynamical equations, their interactions with the environment, and experimental protocols. To preserve high performance when defining new models, most simulators offer two options: low-level programming or description languages. The first option requires expertise, is prone to errors, and is problematic for reproducibility. The second option cannot describe all aspects of a computational experiment, such as the potentially complex logic of a stimulation protocol. Brian addresses these issues using runtime code generation. Scientists write code with simple and concise high-level descriptions, and Brian transforms them into efficient low-level code that can run interleaved with their code. We illustrate this with several challenging examples: a plastic model of the pyloric network, a closed-loop sensorimotor model, a programmatic exploration of a neuron model, and an auditory model with real-time input. Simulating the brain starts with understanding the activity of a single neuron. From there, it quickly gets very complicated. To reconstruct the brain with computers, neuroscientists have to first understand how one brain cell communicates with another using electrical and chemical signals, and then describe these events using code. At this point, neuroscientists can begin to build digital copies of complex neural networks to learn more about how those networks interpret and process information. To do this, computational neuroscientists have developed simulators that take models for how the brain works to simulate neural networks. These simulators need to be able to express many different models, simulate these models accurately, and be relatively easy to use. Unfortunately, simulators that can express a wide range of models tend to require technical expertise from users, or perform poorly; while those capable of simulating models efficiently can only do so for a limited number of models. An approach to increase the range of models simulators can express is to use so-called ‘model description languages’. These languages describe each element within a model and the relationships between them, but only among a limited set of possibilities, which does not include the environment. This is a problem when attempting to simulate the brain, because a brain is precisely supposed to interact with the outside world. Stimberg et al. set out to develop a simulator that allows neuroscientists to express several neural models in a simple way, while preserving high performance, without using model description languages. Instead of describing each element within a specific model, the simulator generates code derived from equations provided in the model. This code is then inserted into the computational experiments. This means that the simulator generates code specific to each model, allowing it to perform well across a range of models. The result, Brian 2, is a neural simulator designed to overcome the rigidity of other simulators while maintaining performance. Stimberg et al. illustrate the performance of Brian 2 with a series of computational experiments, showing how Brian 2 can test unconventional models, and demonstrating how users can extend the code to use Brian 2 beyond its built-in capabilities.
Journal Article
Computational trust models and machine learning
\"This book provides an introduction to computational trust models from a machine learning perspective. After reviewing traditional computational trust models, it discusses a new trend of applying formerly unused machine learning methodologies, such as supervised learning. The application of various learning algorithms, such as linear regression, matrix decomposition, and decision trees, illustrates how to translate the trust modeling problem into a (supervised) learning problem. The book also shows how novel machine learning techniques can improve the accuracy of trust assessment compared to traditional approaches\"-- Provided by publisher.
Metrics for software conceptual models
by
Piattini, Mario
,
Calero, Coral
,
Genero, Marcela
in
Computer software
,
Computer software -- Mathematical models
,
COMPUTERS
2005
The idea that “measuring quality is the key to developing high-quality software systems” is gaining relevance. Moreover, it is widely recognised that the key to obtaining better software systems is to measure the quality characteristics of early artefacts, produced at the conceptual modelling phase. Therefore, improving the quality of conceptual models is a major step towards the improvement of software system development. Since the 1970s, software engineers had been proposing high quantities of metrics for software products, processes and resources but had not been paying any special attention to conceptual modelling. By the mid-1990s, however, the need for metrics for conceptual modelling had emerged. This book provides an overview of the most relevant existing proposals of metrics for conceptual models, covering conceptual models for both products and processes.