Categories
News

Microsoft reveals the hardware needed to run ChatGPT

Microsoft reveals the hardware needed to run ChatGPT

In the fast-paced world of artificial intelligence (AI), having a robust and powerful infrastructure is crucial, especially when you’re working with complex machine learning models like those used in natural language processing. Microsoft Azure is at the forefront of this technological landscape, offering an advanced AI supercomputing platform that’s perfectly suited for the demands of sophisticated AI projects.

At the heart of Azure’s capabilities is its ability to handle the training and inference stages of large language models (LLMs), which can have hundreds of billions of parameters. This level of complexity requires an infrastructure that not only provides immense computational power but also focuses on efficiency and reliability to counter the resource-intensive nature of LLMs and the potential for hardware and network issues.

Azure’s datacenter strength is built on state-of-the-art hardware combined with high-bandwidth networking. This setup is crucial for the effective grouping of GPUs, which are the cornerstone of accelerated computing and are vital for AI tasks. Azure’s infrastructure includes advanced GPU clustering techniques, ensuring that your AI models operate smoothly and efficiently.

What hardware is required to run ChatGPT?

Here are some other articles you may find of interest on the subject of Microsoft Azure :

Software improvements are also a key aspect of Azure’s AI offerings. The platform incorporates frameworks like ONNX, which ensures model compatibility, and DeepSpeed, which optimizes distributed machine learning training. These tools are designed to enhance the performance of AI models while cutting down on the time and resources required for training.

A shining example of Azure’s capabilities is the AI supercomputer built for OpenAI in 2020. This powerhouse system had over 285,000 CPU cores and 10,000 NVIDIA GPUs, using data parallelism to train models on a scale never seen before, demonstrating the potential of Azure’s AI infrastructure.

In terms of networking, Azure excels with its InfiniBand networking, which provides better cost-performance ratios than traditional Ethernet solutions. This high-speed networking technology is essential for handling the large amounts of data involved in complex AI tasks.

Microsoft Azure

Azure continues to innovate, as seen with the introduction of the H100 VM series, which features NVIDIA H100 Tensor Core GPUs. These are specifically designed for scalable, high-performance AI workloads, allowing you to push the boundaries of machine learning.

Another innovative feature is Project Forge, a containerization and global scheduling service that effectively manages Microsoft’s extensive AI workloads. It supports transparent checkpointing and global GPU capacity pooling, which are crucial for efficient job management and resource optimization.

Azure’s AI infrastructure is flexible, supporting a wide range of projects, from small to large, and integrates seamlessly with Azure Machine Learning services. This integration provides a comprehensive toolkit for developing, deploying, and managing AI applications.

In real-world applications, Azure’s AI supercomputing is already making a difference. For instance, Wayve, a leader in autonomous driving technology, uses Azure’s large-scale infrastructure and distributed deep learning capabilities to advance their innovations.

Security is a top priority in AI development, and Azure’s Confidential Computing ensures that sensitive data and intellectual property are protected throughout the AI workload lifecycle. This security feature enables secure collaborations, allowing you to confidently engage in sensitive AI projects.

Looking ahead, Azure’s roadmap includes the deployment of NVIDIA H100 GPUs and making Project Forge more widely available to customers, showing a dedication to continually improving AI workload efficiency.

To take advantage of Azure’s AI capabilities for your own projects, you should start by exploring the GPU-enabled compute options within Azure and using the Azure Machine Learning service. These resources provide a solid foundation for creating and deploying transformative AI applications that can lead to industry breakthroughs and drive innovation.

Image Source : Microsoft

Filed Under: Hardware, Top News





Latest timeswonderful Deals

Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, timeswonderful may earn an affiliate commission. Learn about our Disclosure Policy.

Categories
News

Pinokio AI virtual computer, easily install, run, automate any AI app

Pinokio AI virtual computer makes it easy to automate any AI app

Setting up and running the wealth of artificial intelligence applications, models and tools available can take considerable time. however one tool called Pinokio has been specifically designed to let you install, run and automate any AI application or model with a single click. Allowing you to enjoy using the AI rather than the setup process.

The process of installing, running, and controlling these AI engines can be a daunting task. Pinokio is basically an autonomous virtual computer that simplifies this process by automating command line processes and enabling users to create and share scripts with a single click. “Just like a web browser, Pinokio doesn’t do anything on its own, but will become more and more useful as people build and share apps, workflows, and APIs around it.

Pinokio is a browser that lets you install, run, and automate any AI applications and models automatically and effortlessly. No more opening the terminal. No more git clone. No more conda install. No more pip install. No more messing with execution environments.

The beauty of Pinokio lies in its simplicity. It eliminates the need for opening the terminal, cloning git, installing conda, pip, or dealing with complex execution environments. All these tasks are automated with a single click, making it as user-friendly as a browser. In essence, anything a human can do on a computer, Pinokio can do automatically.

AI applications often require users to open a terminal and enter commands, and sometimes deal with complex environment and installation settings. With Pinokio, all these tasks can be packaged into a simple JSON script, which can then be run in a browser setting with just one click.

How to use Pinokio

Pinokio’s capabilities extend beyond just running commands. It can compose and download files, accumulate data, install libraries and other applications, run shell commands, make network requests, publish files, and browse the internet. Essentially, it can perform pretty much anything a human can do on a computer, without requiring human intervention.

Other articles you may find of interest on the subject of AI automation :

 

One of the key features of Pinokio is its open-source nature. It is 100% open source and free, encouraging a community of developers to contribute to its growth and refinement. This not only ensures the tool remains updated with the latest advancements but also fosters a collaborative environment where innovation thrives.

Pinokio AI virtual computer

Running a server on a computer is not a straightforward task. It requires opening a terminal, running a series of commands to start the server, and keeping the terminal open to keep them running. Pinokio simplifies this process by autonomously reading, writing, processing, and executing anything on your computer, all with a simple scripting language.

Moreover, Pinokio is not limited to a specific operating system. It works seamlessly on Windows, Mac, and Linux, making it a versatile tool for any user. Furthermore, it supports networking, file systems, memory management, data structure, shell execution, and task scheduling, making it a comprehensive solution for managing AI engines.

Now anyone can run powerful server based apps on their own computer, effotlessly:

  • Database Systems: Elasticsearch, MongoDB, RocksDB, Vector Databases, etc.
  • Decentralized Applications: Bitcoin, IPFS, etc.
  • AI Servers: StableDiffusion Web UI, Gradio, Langchain apps, etc.
  • Web apps: Any web apps, really, can be run in the Pinokio browser.
  • Bots: Spin up bots that run in the background, in the Pinokio browser, with one click.

Pinokio is an application that can autonomously read, write, process, and execute anything on your computer, with a simple scripting language. Pinokio can:

  • compose files
  • download files
  • accumulate data
  • install libraries and other applications
  • run shell commands
  • make network requests
  • publish files
  • browse the internet
  • and more.

Pinokio is a game-changer in the world of AI applications. It simplifies the installation and control of AI engines, automates command line processes, and allows the creation and sharing of scripts – all with a single click. As AI continues to evolve, tools like Pinokio will become increasingly invaluable, making complex tasks simpler and more accessible to a wider audience. For more information jump over to the official website or read the comprehensive documentation to get started.

Filed Under: Guides, Top News





Latest timeswonderful Deals

Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, timeswonderful may earn an affiliate commission. Learn about our Disclosure Policy.

Categories
News

How to run AI models on a Raspberry Pi and single board computers (SBC)

running AI models on single board computers SBC

If you are looking for a project to keep you busy this weekend you might be interested to know that it is possible to run artificial intelligence in the form of large language models (LLM) on small single board computers (SBC) such as the Raspberry Pi and others. With the launch of the new Raspberry Pi 5 this month its now possible to carry out more power intensive tasks to its increased performance.

Although before you start it’s worth remembering that running AI models, particularly large language models (LLMs), on a Raspberry Pi or other SBCs presents an interesting blend of challenges and opportunities. While you trade off computational power and convenience, you gain in terms of cost-effectiveness, privacy, and hands-on learning. It’s a field ripe for exploration, and for those willing to navigate its limitations, the potential for innovation is significant.

One of the best ways of accessing ChatGPT from your Raspberry Pi setting up a connection to the OpenAI API, building programs using Python, JavaScript and other programming languages to connect to ChatGPT remotely. Although if you are looking for a a more locally installed more secure version which runs AI directly on your mini PC you will need to select a lightweight LLM that is capable of running and answering your queries more effectively.

Running AI models on a Raspberry Pi

Watch the video below to learn more about how this can be accomplished thanks to Data Slayer if you are interested in learning more about how to utilize the power of your mini PC I deftly recommend you check out his other videos.

Other articles we have written that you may find of interest on the subject of Raspberry Pi 5 :

Before diving in, it’s important to outline the challenges. Running a full-scale LLM on a Raspberry Pi is not as straightforward as running a simple Python script. These challenges are primarily:

  • Limited Hardware Resources: Raspberry Pi offers less computational power compared to typical cloud-based setups.
  • Memory Constraints: RAM can be a bottleneck.
  • Power Consumption: LLMs are known to be energy-hungry.

Benefits of running LLM is on single board computers

Firstly, there’s the compelling advantage of affordability. Deploying AI models on cloud services can accumulate costs over time, especially if you require significant computational power or need to handle large data sets. Running the model on a Raspberry Pi, on the other hand, is substantially cheaper in the long run. Secondly, you gain the benefit of privacy. Your data never leaves your local network, a perk that’s especially valuable for sensitive or proprietary information. Last but not least, there’s the educational aspect. The hands-on experience of setting up the hardware, installing the software, and troubleshooting issues as they arise can be a tremendous learning opportunity.

Drawbacks due to the lack of computational power

However, these benefits come with distinct drawbacks. One major issue is the limited hardware resources of Raspberry Pis and similar SBCs. These devices are not designed to be powerhouses; they lack the robust computational capabilities of a dedicated server or even a high-end personal computer. This limitation is particularly pronounced when it comes to running Large Language Models (LLMs), which are notorious for their appetite for computational resources. Memory is another concern; Raspberry Pis often come with a limited amount of RAM, making it challenging to run data-intensive models. Furthermore, power consumption can escalate quickly, negating some of the cost advantages initially gained by avoiding cloud services.

Setting up your mini PC

Despite these challenges, there have been advancements that make it possible to run LLMs on small computers like Raspberry Pi. One notable example is the work of Georgie Gregov, who ported the Llama model, a collection of private LLMs shared by Facebook, to C++. This reduced the size of the model significantly, making it possible to run on tiny devices like Raspberry Pi.

Running an LLM on a Raspberry Pi is a multi-step process. First, the Ubuntu server is loaded onto the Raspberry Pi. An external drive is then mounted to the Pi, and the model is downloaded to the drive. The next step involves cloning a git repo, compiling it, and moving the model into the repo file. Finally, the LLM is run on the Raspberry Pi. While the process might be a bit slow, it can handle concrete questions well.

It’s important to note that LLMs are still largely proprietary and closed-source. While Facebook has released an open-source version of its Llama model, many others are not publicly available. This can limit the accessibility and widespread use of these models. One notable example is the work of Georgie Gregov, who ported the Llama model, a collection of private LLMs shared by Facebook, to C++. This reduced the size of the model significantly, making it possible to run on tiny devices like Raspberry Pi.

Running AI models on compact platforms like Raspberry Pi and other single-board computers (SBCs) presents a fascinating mix of advantages and limitations. On the positive side, deploying AI locally on such devices is cost-effective in the long run, eliminating the recurring expenses associated with cloud-based services. There’s also an increased level of data privacy, as all computations are carried out within your own local network. Additionally, the hands-on experience of setting up and running these models offers valuable educational insights, especially for those interested in the nitty-gritty of both hardware and software.

However, these benefits come with their own set of challenges. The most glaring issue is the constraint on hardware resources, particularly when attempting to run Large Language Models (LLMs). These models are computational and memory-intensive, and a Raspberry Pi’s limited hardware isn’t built to handle such heavy loads. Power consumption can also become an issue, potentially offsetting some of the initial cost benefits.

In a nutshell, while running AI models on Raspberry Pi and similar platforms is an enticing proposition that offers affordability, privacy, and educational value, it’s not without its hurdles. The limitations in computational power, memory, and energy efficiency can be significant, especially when dealing with larger, more complex models like LLMs. Nevertheless, for those willing to tackle these challenges, the field holds considerable potential for innovation and hands-on learning.

Filed Under: Guides, Top News





Latest timeswonderful Deals

Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, timeswonderful may earn an affiliate commission. Learn about our Disclosure Policy.

Categories
News

LM Studio makes it easy to run AI models locally on your PC, Mac

LM Studio makes it easy to run AI models locally on your PC Mac and Linux

If you are interested in trying out the latest AI models and large language models that have been trained in different ways. Or would simply like one of the open source AI models running locally on your home network. Assisting you with daily tasks. You will be pleased to know that it is really easy to run LLM and hence AI Agents on your local computer without the need for third-party servers. Obviously the more powerful your laptop or desktop computer have the better, but as long as you have 8GB of RAM as a minimum you should be able to run at least one or two smaller AI models such as Mistral and others.

Running AI models locally opens up opportunities for individuals and small businesses to experiment and innovate with AI without the need for expensive servers or cloud-based solutions. Whether you’re a student, an AI enthusiast, or a professional researcher, you can now easily run AI models on your PC, Mac, or Linux machine.

One of the most user-friendly tools for this purpose is LM Studio, a software that allows you to install and use a variety of AI models. With a straightforward installation process, you can have LM Studio set up on your computer in no time. It supports a wide range of operating systems, including Windows, macOS, and Linux, making it accessible to a broad spectrum of users.

The user interface of LM Studio is designed with both beginners and advanced users in mind. The advanced features are neatly tucked away, so they don’t overwhelm new users but are easily accessible for those who need them. For instance, you can customize options and presets to tailor the software to your specific needs.

LM Studio dashboard

LM Studio dashboard chat box

Other articles we have written that you may find of interest on the subject of running AI models locally

LM Studio supports several AI models, including large language models. It even allows for running quantized models in GF format, providing a more efficient way to run these models on your computer. The flexibility to download and add different models is another key feature. Whether you’re interested in NLP, image recognition, or any other AI application, you can find a model that suits your needs.

Search for  AI models and LLMs

LM Studio dashboard

AVX2 support required

Your computer will need to support AVX2 here are a few ways to check what CPU or system is running. Once you know you can then do a quick search to see if the specifications list support for AVX2. You can also ask ChatGPT once you know your CPU.  obviously CPUs after it’s OpenAI’s cut-off date are most likely to support AVX2.

Windows:

  1. Open the Command Prompt.
  2. Run the command systeminfo.
  3. Look for your CPU model in the displayed information, then search for that specific CPU model online to find its specifications.

macOS:

  1. Go to the Apple Menu -> About This Mac -> System Report.
  2. Under “Hardware,” find the “Total Number of Cores” and “Processor Name.”
  3. Search for that specific CPU model online to check its specifications.

Linux:

  1. Open the Terminal.
  2. Run the command lscpu or cat /proc/cpuinfo.
  3. Check for the flag avx2 in the output.

Software Utility:

You can use third-party software like CPU-Z (Windows) or iStat Menus (macOS) to see detailed specifications of your CPU, including AVX2 support.

Vendor Websites:

Visit the CPU manufacturer’s website and look up your CPU model. Detailed specifications should list supported instruction sets.

Direct Hardware Check:

If you have the skill and comfort level to do so, you can directly check the CPU’s markings and then refer to vendor specifications.

For Windows users with an M2 drive, LM Studio can be run on this high-speed storage device, providing enhanced performance. However, as mentioned before, regardless of your operating system, one crucial factor to consider is the RAM requirement. As a rule of thumb, a minimum of 8 GB of RAM is recommended to run smaller AI models such as Mistral. Larger models may require more memory, so it’s important to check the specifications of the models you’re interested in using.

In terms of model configuration and inference parameters, LM Studio offers a range of options. You can tweak these settings to optimize the performance of your models, depending on your specific use case. This level of control allows you to get the most out of your AI models, even when running them on a personal computer.

One of the most powerful features of LM Studio is the ability to create a local host and serve your model through an API. This means you can integrate your model into other applications or services, providing a way to operationalize your AI models. This feature transforms LM Studio from a mere tool for running AI models locally into a platform for building and deploying AI-powered applications.

Running AI models locally on your PC, Mac, or Linux machine is now easier than ever. With tools like LM Studio, you can experiment with different models, customize your settings, and even serve your models through an API. Whether you’re a beginner or a seasoned AI professional, these capabilities open up a world of possibilities for innovation and exploration in the field of AI.

Filed Under: Guides, Top News





Latest timeswonderful Deals

Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, timeswonderful may earn an affiliate commission. Learn about our Disclosure Policy.

Categories
News

How to install Ollama LLM locally to run Llama 2, Code Llama

How to install Ollama locally to run Llama 2 and other LLm models

Large language models (LLMs) have become a cornerstone for various applications, from text generation to code completion. However, running these models locally can be a daunting task, especially for those who are not well-versed in the technicalities of AI.  This is where Ollama comes into play.

Ollama is a user-friendly tool designed to run large language models locally on a computer, making it easier for users to leverage the power of LLMs. This article will provide a comprehensive guide on how to install and use Ollama to run Llama 2, Code Llama, and other LLM models.

Ollama is a tool that supports a variety of AI models including LLaMA-2, uncensored LLaMA, CodeLLaMA, Falcon, Mistral, Vicuna model, WizardCoder, and Wizard uncensored. It is currently compatible with MacOS and Linux, with Windows support expected to be available soon. Ollama operates through the command line on a Mac or Linux machine, making it a versatile tool for those comfortable with terminal-based operations.

Easily install and use Ollama locally

One of the unique features of Ollama is its support for importing GGUF and GGML file formats in the Modelfile. This means if you have a model that is not in the Ollama library, you can create it, iterate on it, and upload it to the Ollama library to share with others when you are ready.

 

 

Installation and Setup of Ollama

To use Ollama, users first need to download it from the official website. After downloading, the installation process is straightforward and similar to other software installations. Once installed, Ollama creates an API where it serves the model, allowing users to interact with the model directly from their local machine.

Downloading and Running Models Using Ollama

Running models using Ollama is a simple process. Users can download and run models using the ‘run’ command in the terminal. If the model is not installed, Ollama will automatically download it first. This feature saves users from the hassle of manually downloading and installing models, making the process more streamlined and user-friendly.

Creating Custom Prompts with Ollama

Ollama also allows users to create custom prompts, adding a layer of personalization to the models. For instance, a user can create a model called ‘Hogwarts’ with a system prompt set to answer as Professor Dumbledore from Harry Potter. This feature opens up a world of possibilities for users to customize their models according to their specific needs and preferences.

Removing Models from Ollama

Just as adding models is easy with Ollama, removing them is equally straightforward. Users can remove models using the ‘remove’ command in the terminal. This feature ensures that users can manage their models efficiently, keeping their local environment clean and organized.

Ollama is a powerful tool that simplifies the process of running large language models locally. Whether you want to run Llama 2, Code Llama, or any other LLM model, Ollama provides a user-friendly platform to do so. With its support for custom prompts and easy model management, Ollama is set to become a go-to tool for AI enthusiasts and professionals alike. As we await the Windows version, Mac and Linux users can start exploring the world of large language models with Ollama.

Filed Under: Guides, Top News





Latest timeswonderful Deals

Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, timeswonderful may earn an affiliate commission. Learn about our Disclosure Policy.

Categories
News

‘Queens On The Run’ Movie Review: Story, Cast and More

Are you looking for a good comedy movie to watch? If so, you should check out Queens on the Run-on Netflix. Queens on the Run is a new Netflix original Spanish film directed by Jorge Macaya and written by Martha Higareda. The film’s plot focuses on four lifelong friends who decide to take the road trip they’ve been planning since high school. However, what begins as a straightforward road trip becomes a wild adventure. Keep reading to find out more about the film.

‘Queens On The Run’: Story

‘Queens on the Run,’ aka ‘Fuga de Renias,’ is the narrative of four friends who are either married or seeking love. Despite being married, they are presently unhappy in their unions. Four friends, dissatisfied with their mediocre lifestyles, embark on a road trip to complete a bucket list they created when they were younger.

On their impromptu voyage, they encounter both fascinating and dangerous individuals. They are confronted with the truth they have been avoiding, and by the film’s conclusion, they emerge as better people than before.

Cast

Martha Higareda portrays Paty Fenix, the trophy wife of an arrogant and self-absorbed assemblyman, Esteban. Alejandra Ambrosi plays the housewife Marilu Davila, who sacrificed her dreams and aspirations for her spouse and children.

Valeria Vera, who portrays Estrella Solares, is the only unmarried quartet member. She is funky, quirky, and still seeking love. Paola Nunez, who plays Famela Guerra, is a workaholic who disregards her spouse and is desperate to become pregnant. Each character is unique, but when they decide together, it’s as if they share a single consciousness.

Queens On The Run Overview

Name of the Movie Queens On The Run
Language Spanish
Director Jorge Macaya
Staring Ricardo Muoz Senior and Paola Nunez
Cast Ricardo Muoz Senior, Paola Nunez, Martha Higareda, Claudia Pineda, Valeria Vera, Alejandra Ambrosi
Released Date 14 April 2023
Film Industry Spanish Film industry
Writer Martha Higareda
Genre Action, Comedy
Queens On The Run

Complete Cast List of “Queens On The Run”

  • Paola Nunez as Famela
  • Martha Higareda as Paty
  • Alejandra Ambrosi as Marilú
  • Valeria Vera as Estrella
  • Claudia Pineda as Lola
  • Ricardo Muñoz Senior as Dr. Claudio
  • Mauricio Isaac
  • Arturo Barba
  • Horacio Garcia Rojas
  • Enrique Arreola
  • Seth Allyn Austin
  • Santiago Michel
  • Miri Higareda

Crew List

Montserrat Carbajal executive post-production producer
Alexis Fridman producer
Julio Cesar Hernandez Vieyra line producer
Martha Higareda producer
Miri Higareda producer
Viridiana Torres supervising producer
Héctor Villegas producer

FAQs

Who is the director of Queens on the Run?

Jorge Macaya was the film’s director.

Who is the writer of “Queens on the Run”?

Martha Higareda wrote the screenplay.

When was “Queens On The Run” released?

Queens On The Run was released on April 14, 2023.