Nueva publicación

查找

Resumen
· 13 mayo, 2024

有奖调研即将截止!您最常用的代码托管平台是什么?

Hi 开发者们,

为了更好地优化开发者社区服务,我们希望了解中国的开发者们最常用的代码托管平台是什么。我们发起了一项有奖调研。调研将于5月14日截止,欢迎参加!

调研时间:2024年4月23日-2024年5月14日

调研方式:

为了便于您的参与,我们准备了两种调研方式,您可以选择任意一种方式参与调研

1. 方式一:点击【此处】进入调研表单,参与调研。

建议您从电脑端参与调研,为了快速完成调研,调研前务必准备好您的社区主页URL。

如何找到社区主页URL:在开发者社区页面点击“个人头像”后,复制浏览器中的URL即可,格式如下(例):https://cn.community.intersystems.com/user/claire-zheng

2. 方式二:登录开发者社区,在这篇帖子下跟帖留言,注明您最常用的代码托管平台即可。

调研奖励:

调研结束后,我们将进行随机抽奖,为参与此次调研的开发者们送出以下奖品——

  • 迷你桌面吸尘器5份
  • 充气颈枕2份
  • 以及特别奖励1份,奖品为:Apple AirTag

快来参加吧!

Resumen
· 13 mayo, 2024

【週間ダイジェスト】 5/06 ~ 5/12 の開発者コミュニティへの投稿

Pregunta
· 13 mayo, 2024

SQL LOAD DATA ISSUES

Hello Community,

I'm trying to load a file by execute this LOAD DATA SQL command in both studio and SQL-SMP I got the below error.

///sample executed queries
ClassMethod ExecutedQueries()
{
    /// executed queries - 1
 LOAD DATA FROM FILE 'C:\\FHIRFILE\mycont.csv' COLUMNS ( C_COUNTRY VARCHAR ( 40 ) , C_CODE VARCHAR ( 20 ) ) INTO SAMPLE . COUNTRIES ( COUNTRY , CODE ) VALUES ( C_COUNTRY , C_CODE )
    /// executed queries - 2
LOAD DATA FROM FILE 'C:\FHIRFILE\mycont.csv' INTO SAMPLE . COUNTRIES ( COUNTRY , CODE )
    
}

sample data

4 comentarios
Comentarios (4)2
Inicie sesión o regístrese para continuar
Artículo
· 12 mayo, 2024 Lectura de 5 min

IRIS AI Studio: A playground to explore the Generative AI & Vector Embedding capabilities

 

Problem

Do you resonate with this - A capability and impact of a technology being truly discovered when it's packaged in a right way to it's audience. Finest example would be, how the Generative AI took off when ChatGPT was put in the public for easy access and not when Transformers/RAG's capabilities were identified. At least a much higher usage came in, when the audience were empowered to explore the possibilities.  

 

Motivation

Recently I got to participate in MIT Grand Hack, Boston where during my conversation with other participants, I noticed immense interest from physicians and non-technical veterans to explore the Generative AI capabilities in their domain. In one of the talks, a finest example, that showcased how the EHR's have a complicated UI being compared with a sleek UI of ChatGPT. Though they both address different needs and comes with it's own pros and cons, my key takeaway was, "The world is not going back towards, complicated user interfaces, or slow systems. And the future lies in systems characterized by intuitive design and optimized performance".

For the non-technical stakeholders who wanted to explore the Generative AI capabilities, is there a way to do it without much of technical hassle. Indeed, Co-lab/Jupyter notebooks have significantly simplified the process of running code. However, can we further streamline this to resemble the intuitive interfaces of platforms like OpenAI's Playground or Azure/Vertex AI Studio? Such a user-friendly environment would empower non-technical stakeholders to grasp the capabilities and constraints of Vector DB effortlessly. 

⚡️ How might we enable non-technical individuals, to evaluate IRIS vector DB potential use cases for their products/services, without exhausting developer resources?

 

Solution 

IRIS AI Studio— A no-code/low-code platform to explore the capabilities of vector embeddings in IRIS DB. User can seamlessly load data from various sources as vector embeddings into IRIS DB and then retrieve the information through multiple channels. Here, the loading of data is called Connectors and retrieval part can be done through Playground. At its core, IRIS AI Studio simplifies the process of working with vector embeddings and let user visualize the power of AI being harnessed for diverse applications.

Connectors

In connectors, the user starts with choosing the data source where they wanted to get the data from to load into IRIS DB as vector embeddings. They can upload files from their local system or retrieve data from cloud storage services as well. This feature has the potential to function as a data pipeline with the addition of CRON job capabilities. 

Initially I have added 4 different data sources - Local Storage, AWS S3, Airtable and Azure Blob Storage.

After choosing the data source, user needs to input the concerned additional information, like client id and secret key for AWS S3 storage. Then, user can configure the embeddings, by choosing the indexing type, model they wanted to use, dimensions and enter the data table name in IRIS, this data needs to be loaded into. Currently indexing from OpenAI and Cohere embeddings are supported, but just with few lines of code, we could add support for many other embeddings. 

After finishing the two steps, the data from the source is converted as vector embeddings and added to the IRIS data table. 

Playground

Now that we have the vector embedded data in IRIS DB, letting user retrieve the contents in their preferred channels can be done from Playground. Here are few examples how that's done. 

In Semantic Search user needs to enter the configurations that they've used to load the data and the semantic search query that they wanted to get executed on the vector embeddings. This will retrieve the contents from IRIS DB and return the results in a natural language format.

 

 

In the Chat with Docs user needs to enter the configurations that they've used to load the data and the chat query they wanted to make. The chat option gets differed from the query option significantly due to the context awareness. Also, in chat with docs we could plug in with many different LLM options to retrieve the contents and need not stick to the chosen indexing options (Currently OpenAI's GPT 3.5 turbo is added as default)

 

Settings

In the settings section, user could add multiple IRIS instance configurations which will be stored in the browser's local storage and user could choose which of them they wanted to actively engage with. This will make their work easier on managing multiple instances and choosing preferred ones to use. Also, here the LLM service's API Keys can be added (will be saved to browser's session storage) and automatically picked up during indexing or retrieval purposes. None of this information is being saved and it's been taken to the backend only for processing purposes.

 

Try it out

Clone the project repository from the following GitHub link: https://github.com/ikram-shah/iris-ai-studio.

Follow the provided instructions to set up the project locally on your machine. Please note, as on May 12, 2024 for some modules, development is still in-progress, but the functionality discussed above should work like a charm 💪🏻 Let know if something doesn’t work as expected through DM or issues section in Github.

Tech Stack

Frontend: VueJS, TailwindCSS, Flowbite

Backend: Python, Flask

Database: InterSystems IRIS

Frameworks/Libraries/Services: Llama-Index, SQLalchemy-iris, OpenAI, Cohere

Infrastructure: Vercel (frontend hosting), Render (backend hosting)

 

Credits

Thanks to the detailed documentation on Vector Search by InterSystems

Thanks to IRIS Vector Search template for simplifying the modules and capabilities by @Alvin Ryanputra 

Thanks to llama-iris library by @Dmitry Maslennikov 

 

Thoughts and Feedback

I come from Product Innovation and Engineering background, and for over a year now, I've been exploring the technological advancements in the healthcare sector. I'm here learning more from the ISC developers community and I am positive that this AI Studio platform would be of help in your work, say developers could share it with Program Managers / Physicians to get their inputs on the quality of data thats been retrieved and many others use cases. If this article resonated with you or sparked any thoughts, please give it a thumbs up 👍 and feel free to share your thoughts in the comments section below. I'd love to connect with anyone interested in discussing further!

🚀 Vote for this application in Vector Search, GenAI and ML contest, if you find it promising!

I will share few more posts that does technical deep dive of this platform. 

6 comentarios
Comentarios (6)1
Inicie sesión o regístrese para continuar