{"id":4722,"date":"2024-04-18T22:49:14","date_gmt":"2024-04-18T22:49:14","guid":{"rendered":"https:\/\/game.intel.com\/?p=4722"},"modified":"2024-05-29T21:16:37","modified_gmt":"2024-05-29T21:16:37","slug":"wield-the-power-of-llms-on-intel-arc-gpus","status":"publish","type":"post","link":"https:\/\/game.intel.com\/gb\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/","title":{"rendered":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs"},"content":{"rendered":"<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<h3 class=\"wp-block-heading\">Easily Run a Variety of LLMs locally with Intel\u00ae Arc\u2122 GPUs<\/h3>\n<\/blockquote>\n\n\n\n<p>Generative AI has changed the landscape of what\u2019s possible in content creation. This technology has the potential to deliver previously unimagined images, videos and writing. Large language models (LLMs) have been making headlines in the age of AI, allowing anyone to prompt their way to generate song lyrics, get answers to complex physics questions, or draft an outline for a slide presentation. And these AI features no longer need to be connected to the cloud or subscription services. They can run locally on your own PC, where you have full control over the model to customize its output.<\/p>\n\n\n\n<p>In this article, we\u2019ll show you how to set up and experiment with popular large language models (LLMs) on a PC with the Intel\u00ae Arc\u2122 A770 16GB graphics card. While this tutorial will make use of the Mistral-7B-Instruct LLM, these same steps can be used with a PyTorch LLM of your choosing such as Phi2, Llama2, etc. And yes, with the latest Llama3 model too!<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">IPEX-LLM<\/h2>\n\n\n\n<p>The reason we can run a variety of models using the same base installation is thanks to <a href=\"https:\/\/github.com\/intel-analytics\/ipex-llm\">IPEX-LLM<\/a>, an LLM library for PyTorch. It is built on top of <a href=\"https:\/\/github.com\/intel\/intel-extension-for-pytorch\">Intel\u00ae Extension for PyTorch<\/a> and contains state-of-art LLM optimizations and low-bit (INT4\/FP4\/INT8\/FP8) weights compression \u2013 with all the latest performance optimizations for Intel hardware. IPEX-LLM takes advantage of the X<sup>e<\/sup>-cores XMX AI acceleration on Intel discrete GPUs like Arc A-series graphics cards for improved performance. It supports Intel Arc A-series graphics on Windows Subsystem for Linux version 2, native Windows environments, and native Linux.<\/p>\n\n\n\n<p>And because all of this is native PyTorch, you can easily swap PyTorch models and input data to run on an Intel Arc GPU with high performance acceleration. This experiment would not have been complete without a performance comparison. Using the instructions below for Intel Arc and commonly available instructions for the competition, we\u2019ve looked at two discrete GPUs positioned in a similar price segment.<\/p>\n\n\n\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69dec2ca835e7&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69dec2ca835e7\" class=\"wp-block-image size-full wp-lightbox-container\"><img fetchpriority=\"high\" width=\"1280\" height=\"720\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-LLM-Execution-on-Arc-A770-2.png\" alt=\"\" class=\"wp-image-4782\"><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewbox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n\n\n\n<p>For example, when running the Mistral 7B model with the IPEX-LLM library, the Arc A770 16GB graphics card can process 70 tokens per second (TPS), or 70% more TPS than the GeForce RTX 4060 8GB using CUDA. What does that mean? A general rule of thumb is that 1 token is equivalent to 0.75 of a word and a good comparison is the <a href=\"https:\/\/wordsrated.com\/speed-reading-statistics\/\">average human reading speed of 4 words per second<\/a> or 5.3 TPS. The Arc A770 16GB graphics card can generate words much faster than the average person can read them!<\/p>\n\n\n\n<p>Our internal testing shows that the Arc A770 16GB graphics card can deliver this capability and competitive or leading performance across a wide range of models compared to the RTX 4060, making Intel Arc graphics a great choice for local LLM execution.<\/p>\n\n\n\n<p>Now let\u2019s go to set up instructions to get you started with LLMs on your Arc A-series GPU.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Installation Instructions<\/h2>\n\n\n\n<p>We can also refer to this page for setting up the environment: <a href=\"https:\/\/ipex-llm.readthedocs.io\/en\/latest\/doc\/LLM\/Quickstart\/install_windows_gpu.html\">Install IPEX-LLM on Windows with Intel GPU \u2014 IPEX-LLM latest documentation<\/a><\/p>\n\n\n\n<p>1. Disable integrated GPU in device manager.<\/p>\n\n\n\n<p>2. Download and install <a href=\"https:\/\/www.anaconda.com\/download\">Anaconda<\/a>.<\/p>\n\n\n\n<p>3. After installation is completed, open the Start menu, search for Anaconda Prompt, run it as administrator, and create a virtual environment using the following commands. Enter each command separately:<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>conda create -n llm python=3.10.6 \n\nconda activate llm\n\nconda install libuv\n\npip install dpcpp-cpp-rt==2024.0.2 mkl-dpcpp==2024.0.0 onednn==2024.0.0 gradio\n\npip install --pre --upgrade ipex-llm&#91;xpu] --extra-index-url https:\/\/pytorch-extension.intel.com\/release-whl\/stable\/xpu\/us\/\n\npip install transformers==4.38.0<\/code><\/pre>\n\n\n\n<p>4. Create a text document named demo.py and save it to C:\\Users\\Your_Username\\Documents or to the directory of your choice.<\/p>\n\n\n\n<p>5. Open demo.py with your favorite editor and copy the following code sample into it:<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>from transformers import AutoTokenizer\nfrom ipex_llm.transformers import AutoModelForCausalLM\nimport torch\nimport intel_extension_for_pytorch\n\ndevice = \"xpu\" # the device to load the model onto\n\nmodel_id = \"mistralai\/Mistral-7B-Instruct-v0.2\" # huggingface model id\n\ntokenizer = AutoTokenizer.from_pretrained(model_id)\nmodel = AutoModelForCausalLM.from_pretrained(model_id, load_in_4bit=True, torch_dtype=torch.float16)\nmodel = model.to(device)\n\nmessages = &#91;\n    {\"role\": \"user\", \"content\": \"What is your favourite condiment?\"},\n    {\"role\": \"assistant\", \"content\": \"Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!\"},\n    {\"role\": \"user\", \"content\": \"Do you have mayonnaise recipes?\"}\n]\n\nencodeds = tokenizer.apply_chat_template(messages, return_tensors=\"pt\")\n\nmodel_inputs = encodeds.to(device)\nmodel.to(device)\n\ngenerated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)\ndecoded = tokenizer.batch_decode(generated_ids)\nprint(decoded&#91;0])<\/code><\/pre>\n\n\n\n<p class=\"has-small-font-size\"><em>Code built from the sample code <a href=\"https:\/\/huggingface.co\/mistralai\/Mistral-7B-Instruct-v0.2\">in this repository<\/a>.<\/em><\/p>\n\n\n\n<p>6. Save demo.py. In Anaconda, navigate to the directory that demo.py is located by using the cd command, and run the following command in the Anaconda prompt:<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>python demo.py<\/code><\/pre>\n\n\n\n<p>Now you can get a nice recipe for making mayonnaise!<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img fetchpriority=\"high\" width=\"1024\" height=\"213\" src=\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-mayo-recipe-1024x213.png\" alt=\"\" class=\"wp-image-4746\"><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">Changing Models<\/h2>\n\n\n\n<p>Using the same environment that we set up above, you can experiment with other popular models on Hugging Face such as llama2-7B-chat-hf, llama3-8B-it, phi-2, gemma-7B-i, and stablelm2 by replacing the Hugging Face model id above in demo.py.<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>model_id = \"mistralai\/Mistral-7B-Instruct-v0.2\" # huggingface model id\n\nto\n\nmodel_id = \"stabilityai\/stablelm-2-zephyr-1_6b\" # huggingface model id<\/code><\/pre>\n\n\n\n<p>Different models might require a different version of transformers package, if you run into errors when launching demo.py, follow the below steps to upgrade\/downgrade transformers:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Open Anaconda Prompt<\/li>\n\n\n\n<li>conda activate llm<\/li>\n\n\n\n<li>pip install transformers==4.37.0<\/li>\n<\/ol>\n\n\n\n<p><strong>Verified transformers versions:<\/strong><\/p>\n\n\n\n<figure class=\"wp-block-table is-style-regular\"><table class=\"has-fixed-layout\"><thead><tr><th class=\"has-text-align-center\" data-align=\"center\">Model ID<\/th><th class=\"has-text-align-center\" data-align=\"center\">Transformers package versions<\/th><\/tr><\/thead><tbody><tr><td class=\"has-text-align-center\" data-align=\"center\">meta-llama\/Llama-2-7b-chat-hf<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.37.0<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">meta-llama\/Meta-Llama-3-8B-Instruct<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.37.0<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">stabilityai\/stablelm-2-zephyr-1_6b<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.38.0<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">mistralai\/Mistral-7B-Instruct-v0.2<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.38.0<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">microsoft\/phi-2<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.38.0<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">google\/gemma-7b-it<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.38.1<\/td><\/tr><tr><td class=\"has-text-align-center\" data-align=\"center\">THUDM\/chatglm3-6b<\/td><td class=\"has-text-align-center\" data-align=\"center\">4.38.0<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<p>Memory requirements may vary per model and framework. For the Intel Arc A750 8GB running with IPEX-LLM, we recommend using Llama-2-7B-chat-hf, Mistral-7B-Instruct-v0.2, phi-2 or chatglm3-6B.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Implementing a ChatBot WebUI<\/h2>\n\n\n\n<p>Now let\u2019s move onto implementing a Gradio chatbot webui for a better experience using your web browser. For more information on implementing an interactive chatbot with LLMs, visit <a href=\"https:\/\/www.gradio.app\/guides\/creating-a-chatbot-fast\">https:\/\/www.gradio.app\/guides\/creating-a-chatbot-fast<\/a><\/p>\n\n\n\n<p>1. Create a document named chatbot_gradio.py in your text editor of choice.<\/p>\n\n\n\n<p>2. Copy and paste the following code snippet into chatbot_gradio.py:<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>import gradio as gr\nimport torch\nimport intel_extension_for_pytorch\nfrom ipex_llm.transformers import AutoModelForCausalLM\nfrom transformers import AutoTokenizer, StoppingCriteria, StoppingCriteriaList, TextIteratorStreamer\nfrom threading import Thread\n\nmodel_id = \"mistralai\/Mistral-7B-Instruct-v0.2\" \n\ntokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)\nmodel = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True, optimize_model=True, load_in_4bit=True, torch_dtype=torch.float16)\nmodel = model.half()\nmodel = model.to(\"xpu\")\nclass StopOnTokens(StoppingCriteria):\n    def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -&gt; bool:\n        stop_ids = &#91;29, 0]\n        for stop_id in stop_ids:\n            if input_ids&#91;0]&#91;-1] == stop_id:\n                return True\n        return False\n\ndef predict(message, history):\n    stop = StopOnTokens()\n    history_format = &#91;]\n    for human, assistant in history:\n        history_format.append({\"role\": \"user\", \"content\": human })\n        history_format.append({\"role\": \"assistant\", \"content\":assistant})\n    history_format.append({\"role\": \"user\", \"content\": message})\n\n    prompt = tokenizer.apply_chat_template(history_format, tokenize=False, add_generation_prompt=True)\n    model_inputs = tokenizer(prompt, return_tensors=\"pt\").to(\"xpu\")\n    streamer = TextIteratorStreamer(tokenizer,  skip_prompt=True, skip_special_tokens=True)\n    generate_kwargs = dict(\n        model_inputs,\n        streamer=streamer,\n        max_new_tokens=300,\n        do_sample=True,\n        top_p=0.95,\n        top_k=20,\n        temperature=0.8,\n        num_beams=1,\n        pad_token_id=tokenizer.eos_token_id,\n        stopping_criteria=StoppingCriteriaList(&#91;stop])\n        )\n    t = Thread(target=model.generate, kwargs=generate_kwargs)\n    t.start()\n\n    partial_message = \"\"\n    for new_token in streamer:\n        if new_token != '&lt;':\n            partial_message += new_token\n            yield partial_message\n\ngr.ChatInterface(predict).launch()<\/code><\/pre>\n\n\n\n<p>3. Open a new anaconda prompt and enter the following commands:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>pip install gradio<\/li>\n\n\n\n<li>conda activate llm<\/li>\n\n\n\n<li>cd to the directory that chat_gradio.py is located in<\/li>\n\n\n\n<li>python chatbot_gradio.py<\/li>\n<\/ul>\n\n\n\n<p>4. Open your web browser and navigate to 127.0.0.1:7860. You should see a chatbot set up with the mistral-7b-instruct-v0.2 language model! You now have a fancy looking webui for your chatbot.<\/p>\n\n\n\n<p>5. Ask a question to start a conversation with your chatbot.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"1469\" height=\"874\" src=\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-chatbot-Q-and-A.png\" alt=\"\" class=\"wp-image-4745\"><\/figure>\n\n\n\n<p><\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity is-style-wide\"\/>\n\n\n\n<p><\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Notices and Disclaimers<\/h3>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<p>Performance varies by use, configuration, and other factors. Learn more on the <a href=\"https:\/\/edc.intel.com\/content\/www\/us\/en\/products\/performance\/benchmarks\/overview\/\">Performance Index site<\/a>.<\/p>\n\n\n\n<p>Performance results are based on testing as of dates shown in configurations and may not reflect all publicly available updates. See backup for configuration details. No product or component can be absolutely secure.<\/p>\n\n\n\n<p>Results that are based on pre-production systems and components as well as results that have been estimated or simulated using an Intel Reference Platform (an internal example new system), internal Intel analysis or architecture simulation or modeling are provided to you for informational purposes only. Results may vary based on future changes to any systems, components, specifications or configurations.<\/p>\n\n\n\n<p>Your costs and results may vary.<\/p>\n\n\n\n<p>Intel technologies may require enabled hardware, software or service activation.<\/p>\n\n\n\n<p>\u00a9 Intel Corporation. Intel, the Intel logo, Arc, and other Intel marks are trademarks of Intel Corporation or its subsidiaries.<\/p>\n\n\n\n<p>*Other names and brands may be claimed as the property of others.<\/p>\n<\/div>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"1280\" height=\"720\" src=\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-System-Configuration-and-Workloads.png\" alt=\"\" class=\"wp-image-4739\" style=\"object-fit:cover\"><\/figure>","protected":false},"excerpt":{"rendered":"<p>Generative AI has changed the landscape of what\u2019s possible in content creation. This technology has the potential to deliver previously unimagined images, videos and writing. Learn how to set up and experiment with popular large language models (LLMs) from the AI community Huggingface on a PC with the Intel\u00ae Arc\u2122 A770 16GB graphics card. <\/p>","protected":false},"author":27,"featured_media":4738,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"inline_featured_image":false,"footnotes":""},"categories":[6],"tags":[45,48,49,14,47],"class_list":["post-4722","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-intel-arc","tag-ai","tag-generative-ai","tag-huggingface","tag-intel-arc-graphics","tag-llms"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v26.8 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/game.intel.com\/gb\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\" \/>\n<meta property=\"og:locale\" content=\"en_GB\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access\" \/>\n<meta property=\"og:description\" content=\"Generative AI has changed the landscape of what\u2019s possible in content creation. This technology has the potential to deliver previously unimagined images, videos and writing. Learn how to set up and experiment with popular large language models (LLMs) from the AI community Huggingface on a PC with the Intel\u00ae Arc\u2122 A770 16GB graphics card.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/game.intel.com\/gb\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\" \/>\n<meta property=\"og:site_name\" content=\"Intel Gaming Access\" \/>\n<meta property=\"article:published_time\" content=\"2024-04-18T22:49:14+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-05-29T21:16:37+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1280\" \/>\n\t<meta property=\"og:image:height\" content=\"768\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Intel Gaming\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@IntelGaming\" \/>\n<meta name=\"twitter:site\" content=\"@IntelGaming\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Intel Gaming\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\"},\"author\":{\"name\":\"Intel Gaming\",\"@id\":\"https:\/\/game.intel.com\/us\/#\/schema\/person\/5a9260725321b6f9dc6b73c2048fb49e\"},\"headline\":\"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs\",\"datePublished\":\"2024-04-18T22:49:14+00:00\",\"dateModified\":\"2024-05-29T21:16:37+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\"},\"wordCount\":1075,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/game.intel.com\/us\/#organization\"},\"image\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png\",\"keywords\":[\"AI\",\"Generative AI\",\"Huggingface\",\"intel arc graphics\",\"LLMs\"],\"articleSection\":[\"Intel\u00ae Arc\u2122 Graphics\"],\"inLanguage\":\"en-GB\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\",\"url\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\",\"name\":\"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access\",\"isPartOf\":{\"@id\":\"https:\/\/game.intel.com\/us\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png\",\"datePublished\":\"2024-04-18T22:49:14+00:00\",\"dateModified\":\"2024-05-29T21:16:37+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#breadcrumb\"},\"inLanguage\":\"en-GB\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-GB\",\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage\",\"url\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png\",\"contentUrl\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png\",\"width\":1280,\"height\":768},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/game.intel.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/game.intel.com\/us\/#website\",\"url\":\"https:\/\/game.intel.com\/us\/\",\"name\":\"Intel Gaming Access\",\"description\":\"Made to Game. Ready for Anything.\",\"publisher\":{\"@id\":\"https:\/\/game.intel.com\/us\/#organization\"},\"alternateName\":\"game.intel.com\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/game.intel.com\/us\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-GB\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/game.intel.com\/us\/#organization\",\"name\":\"Intel Gaming Access\",\"url\":\"https:\/\/game.intel.com\/us\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-GB\",\"@id\":\"https:\/\/game.intel.com\/us\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2026\/01\/square-logo.png\",\"contentUrl\":\"https:\/\/game.intel.com\/wp-content\/uploads\/2026\/01\/square-logo.png\",\"width\":800,\"height\":800,\"caption\":\"Intel Gaming Access\"},\"image\":{\"@id\":\"https:\/\/game.intel.com\/us\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/x.com\/IntelGaming\",\"https:\/\/www.instagram.com\/intelgaming\/\",\"https:\/\/discord.gg\/intel\",\"https:\/\/www.youtube.com\/@intelgaming\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/game.intel.com\/us\/#\/schema\/person\/5a9260725321b6f9dc6b73c2048fb49e\",\"name\":\"Intel Gaming\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-GB\",\"@id\":\"https:\/\/game.intel.com\/us\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/d0fc339c682b4163337309e3b6555e83e4859911e42cdd1109d7b1ddb454cbfb?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/d0fc339c682b4163337309e3b6555e83e4859911e42cdd1109d7b1ddb454cbfb?s=96&d=mm&r=g\",\"caption\":\"Intel Gaming\"},\"url\":\"https:\/\/game.intel.com\/gb\/stories\/author\/caton-lai-intel\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/game.intel.com\/gb\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/","og_locale":"en_GB","og_type":"article","og_title":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access","og_description":"Generative AI has changed the landscape of what\u2019s possible in content creation. This technology has the potential to deliver previously unimagined images, videos and writing. Learn how to set up and experiment with popular large language models (LLMs) from the AI community Huggingface on a PC with the Intel\u00ae Arc\u2122 A770 16GB graphics card.","og_url":"https:\/\/game.intel.com\/gb\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/","og_site_name":"Intel Gaming Access","article_published_time":"2024-04-18T22:49:14+00:00","article_modified_time":"2024-05-29T21:16:37+00:00","og_image":[{"width":1280,"height":768,"url":"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png","type":"image\/png"}],"author":"Intel Gaming","twitter_card":"summary_large_image","twitter_creator":"@IntelGaming","twitter_site":"@IntelGaming","twitter_misc":{"Written by":"Intel Gaming","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#article","isPartOf":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/"},"author":{"name":"Intel Gaming","@id":"https:\/\/game.intel.com\/us\/#\/schema\/person\/5a9260725321b6f9dc6b73c2048fb49e"},"headline":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs","datePublished":"2024-04-18T22:49:14+00:00","dateModified":"2024-05-29T21:16:37+00:00","mainEntityOfPage":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/"},"wordCount":1075,"commentCount":0,"publisher":{"@id":"https:\/\/game.intel.com\/us\/#organization"},"image":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage"},"thumbnailUrl":"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png","keywords":["AI","Generative AI","Huggingface","intel arc graphics","LLMs"],"articleSection":["Intel\u00ae Arc\u2122 Graphics"],"inLanguage":"en-GB","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/","url":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/","name":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs | Intel Gaming Access","isPartOf":{"@id":"https:\/\/game.intel.com\/us\/#website"},"primaryImageOfPage":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage"},"image":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage"},"thumbnailUrl":"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png","datePublished":"2024-04-18T22:49:14+00:00","dateModified":"2024-05-29T21:16:37+00:00","breadcrumb":{"@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#breadcrumb"},"inLanguage":"en-GB","potentialAction":[{"@type":"ReadAction","target":["https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/"]}]},{"@type":"ImageObject","inLanguage":"en-GB","@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#primaryimage","url":"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png","contentUrl":"https:\/\/game.intel.com\/wp-content\/uploads\/2024\/04\/LLM-Blog-041824-llmA770-header.png","width":1280,"height":768},{"@type":"BreadcrumbList","@id":"https:\/\/game.intel.com\/stories\/wield-the-power-of-llms-on-intel-arc-gpus\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/game.intel.com\/"},{"@type":"ListItem","position":2,"name":"Wield The Power of LLMs On Intel\u00ae Arc\u2122 GPUs"}]},{"@type":"WebSite","@id":"https:\/\/game.intel.com\/us\/#website","url":"https:\/\/game.intel.com\/us\/","name":"Intel Gaming Access","description":"Made to Game. Ready for Anything.","publisher":{"@id":"https:\/\/game.intel.com\/us\/#organization"},"alternateName":"game.intel.com","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/game.intel.com\/us\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-GB"},{"@type":"Organization","@id":"https:\/\/game.intel.com\/us\/#organization","name":"Intel Gaming Access","url":"https:\/\/game.intel.com\/us\/","logo":{"@type":"ImageObject","inLanguage":"en-GB","@id":"https:\/\/game.intel.com\/us\/#\/schema\/logo\/image\/","url":"https:\/\/game.intel.com\/wp-content\/uploads\/2026\/01\/square-logo.png","contentUrl":"https:\/\/game.intel.com\/wp-content\/uploads\/2026\/01\/square-logo.png","width":800,"height":800,"caption":"Intel Gaming Access"},"image":{"@id":"https:\/\/game.intel.com\/us\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/IntelGaming","https:\/\/www.instagram.com\/intelgaming\/","https:\/\/discord.gg\/intel","https:\/\/www.youtube.com\/@intelgaming"]},{"@type":"Person","@id":"https:\/\/game.intel.com\/us\/#\/schema\/person\/5a9260725321b6f9dc6b73c2048fb49e","name":"Intel Gaming","image":{"@type":"ImageObject","inLanguage":"en-GB","@id":"https:\/\/game.intel.com\/us\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/d0fc339c682b4163337309e3b6555e83e4859911e42cdd1109d7b1ddb454cbfb?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/d0fc339c682b4163337309e3b6555e83e4859911e42cdd1109d7b1ddb454cbfb?s=96&d=mm&r=g","caption":"Intel Gaming"},"url":"https:\/\/game.intel.com\/gb\/stories\/author\/caton-lai-intel\/"}]}},"_links":{"self":[{"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/posts\/4722","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/users\/27"}],"replies":[{"embeddable":true,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/comments?post=4722"}],"version-history":[{"count":0,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/posts\/4722\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/media\/4738"}],"wp:attachment":[{"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/media?parent=4722"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/categories?post=4722"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/game.intel.com\/gb\/wp-json\/wp\/v2\/tags?post=4722"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}