AI

Google AI Studio 使用方法

今回はGoogleのAI Studioの使用方法を紹介します。

続きを読む

Sora from OpenAI

OpenAI has introduced Sora, an innovative generative AI model designed to transform text into video content. According to OpenAI, Sora can produce 1080p movie-like scenes featuring multiple characters, various types of motion, and detailed background elements, based on either brief or detailed text descriptions or still images. Additionally, Sora has the capability to “extend” existing video clips by filling in missing details.

The model’s proficiency lies in its deep understanding of language, enabling it to interpret prompts accurately and generate dynamic characters that convey vivid emotions. OpenAI emphasizes Sora’s comprehension not only of the user’s requests but also of how those elements manifest in the physical world.

Video Credit : OpenAI

Despite the lofty claims made by OpenAI, the showcased samples from Sora demonstrate impressive capabilities compared to other text-to-video technologies. Sora can produce videos up to a minute long in various styles, such as photorealistic, animated, or black and white, maintaining coherence and avoiding common pitfalls associated with AI-generated content.

However, Sora is not without flaws, as acknowledged by OpenAI. The model may struggle with accurately simulating complex scenes’ physics, understanding cause and effect relationships, or maintaining spatial and temporal consistency. OpenAI positions Sora as a research preview, refraining from making it generally available due to concerns about potential misuse.

OpenAI is actively collaborating with experts to identify and address potential vulnerabilities in the model and is developing tools to detect videos generated by Sora. Should OpenAI decide to make Sora publicly accessible, it pledges to include provenance metadata in generated outputs to mitigate misuse risks.

Yuuma



Google Map with Generative AI feature

Google Maps is set to revolutionize the way users discover new places with the introduction of a generative AI feature, as announced by the company. Leveraging large language models (LLMs), the feature analyzes a vast database of over 250 million locations on Google Maps and taps into contributions from over 300 million Local Guides. By employing a more conversational approach, users can make nuanced queries, such as seeking “places with a vintage vibe in SF,” and Maps will generate tailored recommendations for thrift shops, complete with categories, photo carousels, and review summaries.

Image Credit : Google

Designed to be intuitive, the generative AI feature goes beyond conventional search experiences. Users can engage in a more dynamic interaction by asking follow-up questions. For example, if the initial query relates to vintage shops and is followed by “How about lunch?” the AI will seamlessly transition to finding restaurants that align with the user’s preferences, creating a more personalized and fluid exploration experience.

Image Credit : google

Google emphasizes the inclusivity of the new feature, asserting its capability to generate recommendations for even the most niche or specific queries. Initially launching as an early access experiment in the U.S., the feature will be accessible to select Local Guides—Google’s community contributors—before extending to a wider user base in the near future. While the company hasn’t disclosed the countries that will receive the feature, the move aligns with Google Maps’ evolving role from a navigation tool to a comprehensive platform for discovering and exploring new places.

This generative AI feature follows Google’s recent updates to Maps, which transformed the platform into a more sophisticated search tool, incorporating AI-powered features like photo results and suggesting specific places in response to ambiguous queries. Google envisions this innovation as just the starting point, expressing its commitment to further enhancing Maps with generative AI, solidifying its position as a go-to destination for exploration and discovery.



Samsung Live Translation feature on AI-powered Galaxy S24

At Samsung’s latest launch event, the company unveiled its innovative Live Translation feature, a noteworthy addition to the AI-powered Galaxy S24 smartphone line. The Live Translation feature aims to enhance the calling experience by enabling users to make or receive calls in languages they do not speak, providing real-time audio and on-screen translations. Demonstrated during the Unpacked event, the technology showcased how users could effortlessly make a restaurant reservation in a foreign language, with the conversation seamlessly translated within the native Call app. This feature creates a personalized interpreter experience, emphasizing the convenience it brings to users.

Samsung’s Vice President of Product Management, Drew Blackard, likened the Live Translation feature to having a personal interpreter during phone calls. The technology is designed to support audio and text translations for up to 13 languages, ensuring a versatile and inclusive communication experience. Notably, all translation processes occur on the device itself, preserving the privacy of Samsung owners’ phone calls.

Image Credit : Samsung

Furthermore, the Live Translation feature is designed to remember users’ language settings and the preferred language of each contact, streamlining communication for those engaging in frequent international calls or travel.

As part of this language-focused enhancement, Samsung is extending translation capabilities to the Samsung keyboard. Users can now directly translate words and phrases as they type, offering a dynamic language translation experience during text-based conversations. The keyboard’s language detection feature ensures a seamless translation process, proving useful not only for international communication but also for users comfortable communicating in their native language, even when conversing with others who speak the same language. This comprehensive language integration showcases Samsung’s commitment to leveraging AI for an improved and more accessible user experience.



OpenAI GPT API(14) WebAppでの活用

本記事は前回の「OpenAI GPT API(13) WebAppでの活用」の続きです。

続きを読む


アプリ関連ニュース

お問い合わせはこちら

お問い合わせ・ご相談はお電話、またはお問い合わせフォームよりお受け付けいたしております。

tel. 06-6454-8833(平日 10:00~17:00)

お問い合わせフォーム