As a company dedicated to partner success, we realise how important it is for our clients to fully understand new products and technologies before buying them. This is why Boston regularly publish technical whitepapers, giving customers a broader overview of the latest technologies Boston bring to market.
Please use the links below to view our latest white papers.
Globalisierung und lückenloser Handel: Unsere Welt ist vernetzter, denn je. Eine Pionierrolle nimmt die IT-Branche ein, die längst über lokale Grenzen hinweg vernetzt ist. Andere Branchen sind davon noch weit entfernt. So auch der Finanzhandel. Um dies zu ändern, entwickelte Rubinstein & Schmiedel eine neuartige KI-Software, die den Finanzhandel modernisieren soll. Boston lieferte die technologische Grundlage dafür, darunter NVIDIA GPUs. Eine wichtige Komponente, um die innovative KI-Software mit notwendiger Rechenleistung auszustatten, die sie erfolgreich macht.
Released: 18 January, 2023
Intel® Optane™ Persistent Memory (PMem) 200 is the second generation of Intel’s DC Persistent Memory (DCPMM) which is an emerging technology where non-volatile media is placed onto a Dual In-Line Memory Module (DIMM) and installed on the memory bus, traditionally used only for volatile memory. Read our whitepaper to find out what is Intel Optane PMem and why you need it...
Released: 13 July, 2022
finitia verhilft Architektur-, Planungs- und Ingenieurbüros zum Virtualisierungsdurchbruch: Dafür entwickelte finitia von 2019 an eine Virtualisierungslösung über VMware. Boston begleitete seinen Partner vom ersten Briefing an, um die Planungssicherheit und Nachhaltigkeit für Architekten, Planer und Ingenieure in ihrer CAD-Lösung zu erhöhen.
Released: 27 May, 2022
Whether your applications pull terabytes of data from disparate databases or you are building data center infrastructure to accommodate immense scale, planning and implementing a successful flash-first strategy is imperative. This guide explores the feature differences between the Micron 7300 and 7400 SSDs and help you decide which of these SSDs is the best fit for your workloads.
Released: 03 November, 2021
Graphcore, the U.K.-based startup that launched the Intelligence Processing Unit (IPU) for AI acceleration in 2018, has introduced the IPU-Machine. This second-generation platform has greater processing power, more memory and built-in scalability for handling extremely large parallel processing workloads. The well-funded startup has a blue-ribbon pedigree of engineers, advisers and investors, and enjoys a valuation approaching $2 billion. The new MK2 part, manufactured by TSMC, is a massively parallel 59.4 B transistor processor. It delivers some 250 Trillion Operations per Second (TOPS) across 1,472 cores and 900MB of In-Processor Memory interconnected across a 2.8Tb/s low-latency fabric.
Released: 06 April, 2021
The Weka AI™ reference architecture, powered by NVIDIA DGX A100 systems and Weka's industry leading file system WekaFS™, was developed and verified by Weka and NVIDIA.
Released: 15 January, 2021
Early computers integrated computing, storage, and networking resources in a single system. As the need for capacity grew, these elements were disaggregated into separate fiefdoms within the IT infrastructure, making systems more capable but harder to manage. Over the past decade, a few pioneers have pursued the concept of re-integrating these resources into a single system that is more capable and easier to manage. This paper explores that concept, now known as Hyper-Converged Infrastructure (HCI), and demonstrates how it can make IT operations more agile while reducing overall expenses
Released: 29 October, 2020
Built on the brand new NVIDIA A100 Tensor Core GPU, NVIDIA DGX™ A100 is the third generation of DGX systems. Featuring 5 petaFLOPS of AI performance, DGX A100 excels on all AI workloads– analytics, training, and inference–allowing organizations to standardize on a single system that can speed through any type of AI task and dynamically adjust to changing compute needs over time. And with the fastest I/O architecture of any DGX system, NVIDIA DGX A100 is the foundational building block for large AI clusters such as NVIDIA DGX SuperPOD, the enterprise blueprint for scalable AI infrastructure that can scale to hundreds or thousands of nodes to meet the biggest challenges. This unmatched flexibility reduces costs, increases scalability, and makes DGX A100 the universal system for AI infrastructure. In this white paper, we’ll take a look at the design and architecture of DGX A100.
Released: 05 August, 2020
The convergence of Artificial Intelligence (AI) and High Performance Computing (HPC) has been a driving factor of broader adoption of HPC by a wide variety of industries. The time is ripe for AI, and organisations looking to gain an edge in business are turning more and more to AI development to build their next generation of products and services.
Released: 03 February, 2020
Formula 1’s CFD restriction regime has been shaken up big time as the FIA looks to cut the costs of aerodynamic development.
Released: 02 January, 2020
The faster the processor is, the hotter it tends to be when in use. Computer cooling is required to remove the excess heat produced by computer components, to keep them within safe operating temperatures - find out about how liquid cooling has evolved in our white paper...
Released: 23 June, 2019
Build your supercomputing infrastructure on Boston’s extensive industry and design expertise for High Performance Computing applications.
Released: 04 April, 2019
The convergence of Artificial Intelligence (AI) and High Performance Computing (HPC) has been a driving factor of broader adoption of HPC by a wide variety of industries.
Released: 19 March, 2019
Across an evolving digital world, disruptive and emerging technology trends in business, industry, science, and entertainment increasingly impact the world's economies. By 2020, the success of half the world's Global 2000 companies will depend on their abilities to create digitally enhanced products, services, and experiences, and large organisations expect to see an 80 percent increase in their digital revenues, all driven by advancements in technology and usage models they enable.
Released: 11 November, 2018
The performance, scalability and ease of manageability provided by Intel EE for Lustre software make it an excellent choice for entry-level HPC workloads, compared to NFS. The small form factor architecture can perfectly fit the workloads, growing as the workload grows, with the added benefits of excellent performance and manageability.
Released: 25 April, 2017
Originally released as a version 1 specification by the NVM Express Work Group in March 2011 and through several minor revisions since, NVMe has gone from conception to maturity and is now tipped to be one of the most important storage technologies for both server and client computing in the coming decade and beyond.
Released: 24 March, 2017
It's that time of year again when Intel finally release their latest enterprise processor for the dual processor segment to the eagerly awaiting professional market. Following on in the traditional early September launch time frame, the Xeon E5-2600 v3 processor series, codename Haswell-EP has been officially launched, finally allowing us at Bostonlabs to go through all the exciting details of the processors which we’ve been testing secretly in our labs for some time.
Released: 30 September, 2014
This guide, produced by Citrix India, provides high-level design details that describes the architecture for XenDesktop 7.1 Citrix Validated Solution running on Commodity Hardware by Boston. The architecture is based on the fundamentals of how cloud computing works in conjunction with commodity hardware to considerably lower the total cost of ownership for a Citrix XenDesktop environment. This guide has been created through architectural design best practices obtained from Citrix Consulting Services and through lab testing, and is intended to provide guidance for solution evaluation.
Released: 07 July, 2014
By partnering with Citrix India and leveraging commodity hardware by Supermicro, Boston Limited has designed and validated a cost-effective yet robust virtual desktop infrastructure that is scalable from 500 – 10,000 users. Following on from the launch of our VDI solution at CeBIT 2014 (below), Boston has developed another Citrix XenDesktop7-based VDI solution to explore the performance benefits of using class-leading Intel Xeon E5-2600 V2 series processors.
Released: 24 April, 2014
An increasing number of enterprises are looking towards desktop virtualization to help them respond to rising IT costs, security concerns, the user demands of BYOD and mobile working strategies. But can a desktop virtualization solution have a lower or equivalent Total Cost of Ownership (TCO) when compared to the traditional approach of procuring physical desktops? Is there a solution which delivers all the benefits of scalability and performance while maintaining a lower TCO? – Leveraging commodity hardware is the answer.
Released: 10 March, 2014
At Boston we believe we are at the inflection point within the supercomputing industry. There are powerful economic drivers disrupting the dominance of the x86 server and just as old vector supercomputers were replaced by x86 commodity supercomputers, the ARM platform looks set to become the next disruptive technology in the commodity chain.
Released: 19 November, 2012
LSI® MegaRAID® CacheCade® Pro 2.0 software promises to be the answer to IT managers' serious pain points with storage, those related to random I/O performance. This new functionality is supported on several LSI SAS 6Gb/s RAID controller lines and adds SSD caching functionality to the already impressive feature set to help improve performance and negate some of the pitfalls of traditional hard disk technology. Today it is possible to purchase magnetic hard disks with capacities of up to 3TB for under £300 but they struggle to provide 200 IOPS, while some SSDs easily achieve 50,000 IOPs but cost 5-10 times that for only a few hundred gigabytes of capacity.
Released: 25 June, 2012
Web giants such as Google and Facebook are inching towards the Arctic Circle, building their latest data centres in countries such as Finland and Sweden to cope with the exponential demand for their Internet services. What was forbidden terrain for agriculture and manufacturing is now home to power hungry server farms that not only need sustainable sources of energy, but an extremely cold climate to chill servers. When Facebook announced that it was building a data centre in Lulea, Sweden which is less than 100km south of the Arctic Circle, the Guardian reported that 'each of Facebook's US data centres is estimated to use the same amount of electricity as 30,000 US homes. Energy consumption of warehouses run by companies such as Facebook, Google and Amazon, is among the fastest growing sources of global electricity demand.
Released: 15 June, 2012
To help our clients make informed decisions about new technologies, we have opened up our research & development facilities and actively encourage customers to try the latest platforms using their own tools and if necessary together with their existing hardware. Remote access is also available
A summary of what to expect from the RUN:AI event