share_log

Amazon Bedrock General Manager: A Rich Mix of Large Models Should Be Offered for Customers

Amazon Bedrock General Manager: A Rich Mix of Large Models Should Be Offered for Customers

亞馬遜 Bedrock 總經理:應爲客戶提供豐富的大型號組合
鈦媒體 ·  04/25 20:01

AsianFin-- Many cloud computing companies are actively engaged in developing their in-house trained foundational large models. This approach is understandable from a business perspective, but it may not be fully accepted by customers as the choice of large models may be restricted.

AsianFin——許多雲計算公司都在積極參與開發經過內部培訓的基礎大型模型。從業務角度來看,這種方法是可以理解的,但由於大型模型的選擇可能會受到限制,因此可能無法被客戶完全接受。

On one hand, the innovation of large models is yet to reach its peak, and the capabilities of different model providers vary. On the other hand, it's also related to that customer demand scenarios as no single model can meet all scenarios. Therefore, for different use cases, customers need more than one or two models to satisfy all kinds of scenario requirements.

一方面,大型模型的創新尚未達到頂峯,不同模型提供商的能力也各不相同。另一方面,它也與客戶需求場景有關,因爲沒有一個單一的模型可以滿足所有場景。因此,對於不同的用例,客戶需要一個或兩個以上的模型來滿足各種場景需求。

In the past, more than 90% of Amazon Web Services (AWS) products were derived from customer needs. AWS's generative artificial intelligence (AI) strategy also follows this path.

過去,超過90%的亞馬遜網絡服務(AWS)產品源自客戶需求。AWS 的生成式人工智能 (AI) 策略也遵循這條道路。

AWS has also released its own foundational large model, Amazon Titan, in April 2023. This stems from AWS's accumulation of AI technology, such as the widely known voice assistant Alexa, drone delivery service Prime Air, and cashier-less stores Amazon Go, all of which employ a large number of speech, semantic, and visual machine learning technologies.

AWS 還在 2023 年 4 月發佈了自己的基礎大型模型——亞馬遜泰坦。這源於AWS積累的人工智能技術,例如廣爲人知的語音助手Alexa、無人機送貨服務Prime Air和無收銀員商店Amazon Go,所有這些都採用了大量的語音、語義和視覺機器學習技術。

Atul Deo, General Manager of Amazon Bedrock, pointed out that if AWS doesn't have its own model, it means it must rely entirely on partners. Starting from scratch to build models also provides a "hands-on" approach to solving problems.

亞馬遜 Bedrock 總經理 Atul Deo 指出,如果AWS沒有自己的模型,那就意味着它必須完全依賴合作伙伴。從頭開始構建模型還提供了一種 “動手操作” 的方法來解決問題。

As a result, there is an interesting phenomenon: because Amazon Bedrock provides a range of capabilities needed for enterprises to build generative AI applications, it can simplify development while ensuring privacy and security. On Amazon Bedrock, customers can find Amazon Titan as well as current mainstream large model versions, including models from Anthropic, Stability AI, AI21 Labs, Meta, Cohere, and Mixtral... This list goes on and on.

因此,有一個有趣的現象:由於Amazon Bedrock提供了企業構建生成式人工智能應用程序所需的一系列功能,因此它可以在確保隱私和安全的同時簡化開發。在亞馬遜 Bedrock 上,客戶可以找到亞馬遜泰坦以及當前的主流大型模型版本,包括來自 Anthropic、Stability AI、AI21 Labs、Meta、Cohere 和 Mixtral 的模型...這個清單不勝枚舉。

On Tuesday evening, AWS announced multiple feature updates for Amazon Bedrock, which overall enhance efficiency in developing generative AI applications for customers.

週二晚上,AWS宣佈了對Amazon Bedrock的多項功能更新,這總體上提高了爲客戶開發生成式人工智能應用程序的效率。

In addition to feature updates, AWS also provides a range of new models on Amazon Bedrock, including the officially available Amazon Titan Image Generator for image generation, Meta Llama 3, and the preview version of Amazon Titan Text Embeddings V2. Three models from Cohere, Command R and Command R+, are also set to be released.

除了功能更新外,AWS還在亞馬遜Bedrock上提供了一系列新模型,包括正式推出的用於圖像生成的亞馬遜泰坦圖像生成器、Meta Llama 3和亞馬遜泰坦文本嵌入V2的預覽版。Cohere的三款模型,即Command R和Command R+,也將發佈。

In particular, the preview version of Amazon Titan Text Embeddings V2 is optimized for applications such as information retrieval, question-answering chatbots, and personalized recommendations that use RAG (retrieval-augmented generation) technology. Many enterprises adopt RAG technology to enhance the results generated by base models by connecting to knowledge sources, but the issue is that running these operations can consume a lot of computing and storage resources. Amazon Titan Text Embeddings V2 reduces storage and computing costs while maintaining the accuracy of using RAG retrieval results.

特別是,Amazon Titan Text Embeddings V2 的預覽版針對信息檢索、問答聊天機器人和使用 RAG(檢索增強生成)技術的個性化推薦等應用程序進行了優化。許多企業採用RAG技術,通過連接到知識源來增強基礎模型生成的結果,但問題是運行這些操作會消耗大量的計算和存儲資源。亞馬遜 Titan Text Embeddings V2 降低了存儲和計算成本,同時保持了使用 RAG 檢索結果的準確性。

Generative AI requires not only large models but also support from acceleration chips, databases, data analysis, and data security services. From the bottom layer of acceleration chips and storage optimization to the middle layer of model construction tools and services, and finally to the top layer of generative AI-related applications, it can be seen that AWS is attempting to provide an end-to-end technology stack for customers to build generative AI.

生成式 AI 不僅需要大型模型,還需要加速芯片、數據庫、數據分析和數據安全服務的支持。從加速芯片和存儲優化的底層到模型構造工具和服務的中間層,最後到生成式人工智能相關應用程序的頂層,可以看出,AWS正在嘗試爲客戶提供端到端的技術堆棧來構建生成式人工智能。

On the eve of the release, Atul shared with TMTPost his views on generative AI, technical methodology, and how Amazon Bedrock aids customer success.

在發佈前夕,阿圖爾與TMTPost分享了他對生成式人工智能、技術方法以及亞馬遜基岩如何幫助客戶成功的看法。

The following is the transcript of the dialogue, edited by TMTPost for clarity and brevity:

以下是對話記錄,爲清晰和簡潔起見,由 TMTPost 編輯:

TMTPost: What are the different advantages between large companies and small, focused teams in achieving AI technology innovation and industry empowerment?

TMTPost:在實現人工智能技術創新和行業賦能方面,大公司和小型專注團隊有哪些不同的優勢?

Atul: Regarding application deployment for customers, I don't think there are any significant differences between large companies and small businesses; they have many similarities. We all want to try different models for large companies. Currently, Data Hygiene is a demanding job. When it comes to deploying applications for smaller companies, managing and ensuring the high quality and consistency of private data required for model training is relatively easy. But for larger companies, with a large amount of differentiated data that is more dispersed, managing data will be more difficult. On the other hand, startups can act faster as they are less risk-averse. They don't have an existing customer base like large customers, so they may make mistakes and improve quickly through trial and error.

Atul:關於客戶的應用程序部署,我認爲大公司和小型企業之間沒有任何顯著區別;它們有很多相似之處。我們都想爲大公司嘗試不同的模式。當前,數據衛生是一項艱鉅的工作。在爲小型公司部署應用程序時,管理和確保模型訓練所需的私有數據的高質量和一致性相對容易。但是對於大型公司來說,由於大量差異化數據更加分散,管理數據將更加困難。另一方面,初創企業可以更快地採取行動,因爲它們規避風險的程度較低。他們沒有像大客戶那樣現有的客戶群,因此他們可能會犯錯誤並通過反覆試驗快速改進。

TMTPost: What problem does AWS want to address with generative AI?

TMTPost:AWS 想用生成式 AI 解決什麼問題?

Atul: We are actively exploring new possibilities. Whether customers want to build models themselves or customize existing models deeply, we hope to build a generative AI stack that allows customers to use rich and first-class tools. In addition to Amazon SageMaker and rich instance types provided by NVIDIA, we are also actively developing custom chips covering training and inference domains to meet more refined needs.

Atul:我們正在積極探索新的可能性。無論客戶是想自己構建模型還是深度定製現有模型,我們都希望構建一個生成式 AI 堆棧,讓客戶能夠使用豐富的一流工具。除了 Amazon SageMaker 和 NVIDIA 提供的豐富實例類型外,我們還積極開發涵蓋訓練和推理領域的自定義芯片,以滿足更精細的需求。

Through a series of innovations from the bottom layer to the middle layer, our goal is to allow any developer in the enterprise to freely build generative AI applications without worrying about complex machine learning or underlying infrastructure. We firmly believe that the tools provided will be industry-leading and help them achieve innovation breakthroughs in applications.

通過從底層到中間層的一系列創新,我們的目標是允許企業中的任何開發人員自由構建生成式 AI 應用程序,而不必擔心複雜的機器學習或底層基礎架構。我們堅信,所提供的工具將處於行業領先地位,並幫助他們在應用方面實現創新突破。

Currently, we have launched two versions of Amazon Q: Amazon Q business and Amazon Q developer. Amazon Q business aims to equip every employee in the enterprise with a professional consultant to ensure they can quickly get answers and efficiently complete tasks; while Amazon Q developer focuses on improving the efficiency of developers, providing them with instant answers to smoothly complete their specific tasks. This is the ultimate goal of Amazon Q and the direction we are tirelessly pursuing.

目前,我們已經推出了兩個版本的亞馬遜 Q:亞馬遜 Q 企業版和亞馬遜 Q 開發者。Amazon Q business 旨在爲企業中的每位員工配備專業顧問,確保他們能夠快速獲得答案並高效完成任務;而Amazon Q 開發人員則專注於提高開發人員的效率,爲他們提供即時答案以順利完成其特定任務。這是 Amazon Q 的最終目標,也是我們孜孜不倦地追求的方向。

TMTPost: How long will it take for AWS truly change its product and business structure? How to establish AWS's leadership in this field?

TMTPost:AWS 真正改變其產品和業務結構需要多長時間?如何確立 AWS 在這一領域的領導地位?

Atul: Actually, everything depends on customers and the specific problems we are trying to solve. We have seen tens of thousands of customers using SageMaker to change their customer experiences. Some changes have already happened, while others will take some time. Therefore, there is no fixed answer as to when significant changes can be expected.

Atul:實際上,一切都取決於客戶和我們想要解決的具體問題。我們已經看到成千上萬的客戶使用SageMaker來改變他們的客戶體驗。一些變化已經發生,而另一些則需要一些時間。因此,對於何時可以預期會發生重大變化,沒有固定的答案。

For example, the New York Stock Exchange is using Bedrock to analyze and process numerous regulatory files and transform complex regulatory content into easy-to-understand language, which will have a profound impact on end-users; meanwhile, electronic health record technology provider Netsmart has successfully reduced the time for managing patient health records by 50% through the application of relevant technologies, undoubtedly freeing up more time for doctors to care for more patients.

例如,紐約證券交易所正在使用Bedrock來分析和處理大量監管文件,並將複雜的監管內容轉換爲易於理解的語言,這將對最終用戶產生深遠影響;同時,電子健康記錄技術提供商Netsmart通過應用相關技術,成功地將管理患者健康記錄的時間減少了50%,這無疑爲醫生騰出了更多時間來照顧更多患者。

Today we have seen some positive impacts on end-users, but I believe it is still a process that needs time to gradually develop and popularize. However, the pace of progress is relatively fast and the momentum has been building up. Therefore, I cannot predict with certainty whether generative artificial intelligence will become very common by the end of this year or next year. However, what can be certain is that it is gradually changing our world, bringing more convenience and possibilities.

今天,我們已經看到了對最終用戶的一些積極影響,但我相信這仍然是一個需要時間才能逐步發展和普及的過程。但是,進展速度相對較快,勢頭一直在增強。因此,我無法確定生成式人工智能是否會在今年年底或明年變得非常普遍。但是,可以肯定的是,它正在逐漸改變我們的世界,帶來更多的便利和可能性。

TMTPost: For example, RAG is used to solve hallucination problems, but some papers have mentioned that RAG alone cannot solve hallucinations. In enterprise-level applications, how to assess the degree of hallucination and its impact when specific applications are used?

TMTPost:例如,RAG用於解決幻覺問題,但是一些論文提到,光靠RAG無法解決幻覺。在企業級應用程序中,如何評估使用特定應用程序時的幻覺程度及其影響?

Atul: Although we cannot completely eliminate this problem, I believe that more and more cutting-edge research will come up to help with this issue. You will see customers make more progress and improvements in dealing with hallucinations. I can tell you clearly that although this problem cannot be completely solved, it does help reduce its impact and cannot be completely eliminated as part of our action.

阿圖爾:儘管我們無法完全消除這個問題,但我相信越來越多的前沿研究將幫助解決這個問題。您將看到客戶在應對幻覺方面取得更多進展和改進。我可以清楚地告訴你,儘管這個問題無法完全解決,但它確實有助於減少其影響,而且不能作爲我們行動的一部分完全消除。

TMTPost: Regarding the collaboration issue between models, what are AWS's better solutions for customers when multiple models are used?

TMTPost:關於模型之間的協作問題,當使用多個模型時,AWS爲客戶提供的更好的解決方案是什麼?

Atul: This is important for customers. In this regard, we specially launched a feature called model evaluation, which was released in December last year and is planned to be fully launched tomorrow. Essentially, this feature is designed to help customers compare the performance of different models on a given set of prompts so that they can choose the model that best suits their specific use cases.

Atul:這對客戶很重要。在這方面,我們特別推出了一項名爲模型評估的功能,該功能於去年12月發佈,計劃明天全面上線。本質上,此功能旨在幫助客戶比較不同模型在給定提示集上的性能,以便他們可以選擇最適合其特定用例的模型。

To achieve this goal, customers have three options to choose from. First, they can compare the performance of different models based on given prompts in the console; second, customers can use the automated evaluation feature to run different models on different datasets or use standard industry datasets to see which models perform well; finally, customers can also use their internal professional teams to evaluate models in different ways and determine which model meets their expectations. Ultimately, customers will receive a detailed report from Bedrock, which will show the performance of the models and how to decide which models make sense for them.

爲了實現這一目標,客戶有三種選擇可供選擇。首先,他們可以根據控制檯中給定的提示比較不同模型的性能;其次,客戶可以使用自動評估功能在不同的數據集上運行不同的模型,或者使用標準的行業數據集來查看哪些模型表現良好;最後,客戶還可以使用其內部專業團隊以不同的方式評估模型並確定哪種模型符合他們的預期。最終,客戶將收到來自Bedrock的詳細報告,該報告將顯示模型的性能以及如何決定哪些型號對他們有意義。

TMTPost: What initiatives has AWS taken in AI ethics?

TMTPost:AWS 在人工智能倫理方面採取了哪些舉措?

Atul: We are working closely with multiple government organizations. Take our Titan Image Generator, for example. This tool has watermarking capabilities to add invisible watermarks to help customers determine if the generated images are generated by artificial intelligence. In addition, we are also cooperating with a series of other organizations to ensure the responsible use of artificial intelligence.

阿圖爾:我們正在與多個政府組織密切合作。以我們的 Titan 圖像生成器爲例。該工具具有水印功能,可以添加不可見的水印,以幫助客戶確定生成的圖像是否由人工智能生成。此外,我們還與一系列其他組織合作,以確保負責任地使用人工智能。

TMTPost: What is AWS's experience in self-developed chips?

TMTPost:AWS 在自主開發芯片方面的經驗如何?

Atul: Over the years, we have been investing in chip development and acquired chip design company Annapurna Labs as early as 2015. Although our initial focus was on virtualization and general-purpose computing chips, we later focused on developing AI chips specifically for machine learning. For example, two dedicated chips for artificial intelligence training and inference, Amazon Trainium and Amazon Inferentia.

阿圖爾:多年來,我們一直在投資芯片開發,早在2015年就收購了芯片設計公司Annapurna Labs。儘管我們最初的重點是虛擬化和通用計算芯片,但後來我們專注於開發專門用於機器學習的人工智能芯片。例如,兩個用於人工智能訓練和推理的專用芯片,亞馬遜Trainium和亞馬遜Inferentia。

Thanks to years of continuous investment in chip development, we have more opportunities to iterate and improve these chips to ensure their performance and stability. These improvements come at the right time because the demand for computing power in generative AI is growing.

由於多年來對芯片開發的持續投資,我們有更多的機會對這些芯片進行迭代和改進,以確保其性能和穩定性。這些改進恰逢其時,因爲生成式人工智能對計算能力的需求正在增長。

TMTPost: There are many models on Bedrock. Have you observed which model is most popular among customers, such as Meta and Anthropic?

TMTPost:基岩上有很多模型。你有沒有觀察到哪種型號最受客戶歡迎,比如 Meta 和 Anthropic?

Atul: Currently, we will not disclose the specific performance of various model providers. But what I want to say is that these models are favored by a large number of users. This is mainly because the choice of models depends on specific application scenarios, and people will choose different models according to different needs. Therefore, it is too early to identify which models are widely used.

Atul:目前,我們不會透露各種模型提供商的具體表現。但是我想說的是,這些模型受到了大量用戶的青睞。這主要是因爲模型的選擇取決於特定的應用場景,人們會根據不同的需求選擇不同的模型。因此,現在確定哪些模型被廣泛使用還爲時過早。

譯文內容由第三人軟體翻譯。


以上內容僅用作資訊或教育之目的,不構成與富途相關的任何投資建議。富途竭力但無法保證上述全部內容的真實性、準確性和原創性。
    搶先評論