POST/v1/sam-img2img
1const axios = require('axios');
2
3const fs = require('fs');
4const path = require('path');
5
6async function toB64(imgPath) {
7    const data = fs.readFileSync(path.resolve(imgPath));
8    return Buffer.from(data).toString('base64');
9}
10
11const api_key = "YOUR API-KEY";
12const url = "https://api.segmind.com/v1/sam-img2img";
13
14const data = {
15  "image": "toB64('https://segmind.com/kitchen.jpg')"
16};
17
18(async function() {
19    try {
20        const response = await axios.post(url, data, { headers: { 'x-api-key': api_key } });
21        console.log(response.data);
22    } catch (error) {
23        console.error('Error:', error.response.data);
24    }
25})();
RESPONSE
image/jpeg
HTTP Response Codes
200 - OKImage Generated
401 - UnauthorizedUser authentication failed
404 - Not FoundThe requested URL does not exist
405 - Method Not AllowedThe requested HTTP method is not allowed
406 - Not AcceptableNot enough credits
500 - Server ErrorServer had some issue with processing
Expand

Attributes


imageimage *

Image to Segment

To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.

Segment Anything Model (SAM)

The Segment Anything Model (SAM) is the latest groundbreaking AI model introduced by Meta AI. With the power to "cut out" any object from any image with just a single click, SAM is revolutionizing the world of image processing. Zero-shot capability; Unlike other models that require extensive training for each new object or image, SAM boasts of a promptable segmentation system. This means it can easily generalize to unfamiliar objects and images without the need for any additional training.

The SAM model architecture integrates a masked auto-encoder (MAE) and a Vision Transformer (ViT) to generate distinct image embeddings, typically initialized before model prompting. SAM's prompt encoding component transforms elements such as points, masks, and texts into immediate vectors, categorizing prompts into sparse or dense groups. Subsequently, SAM's mask decoder predicts segmentation masks by leveraging these embeddings, aligning them with a specific output token. Through bidirectional cross-attention, the labeled masks are processed to enhance SAM's weights, ensuring its adaptability and streamlined performance.

Advantages

  1. Zero-shot Learning: SAM can identify and segment unfamiliar objects without any prior training, making it highly versatile.

  2. Promptable System: Users can easily guide SAM using prompts, ensuring more accurate results tailored to specific needs.

  3. Speed and Efficiency: Thanks to its advanced architecture, SAM processes images rapidly, delivering results in real-time.

  4. High Precision : SAM's extensive training and hybrid architecture ensure that the segmented results are of the highest quality.

Use cases

  1. E-commerce Product Listing: Retailers can use SAM to automatically segment product images, removing backgrounds and placing products on a consistent backdrop.

  2. Photo Editing Apps: SAM can be integrated into photo editing software, allowing users to easily extract objects and create composite images.

  3. Augmented Reality (AR) Applications: SAM can help in real-time object segmentation, enabling developers to overlay virtual objects onto real-world scenes seamlessly.

  4. Medical Imaging: Healthcare professionals can use SAM to segment specific organs or anomalies in medical images, aiding in diagnosis and treatment planning.

  5. Video Production: Film and video editors can utilize SAM to segment and replace backgrounds, create special effects, or integrate CGI elements into real footage.

SAM License

The Segment Anything Model (SAM) is open source. Meta AI has released the SA-1B Dataset for research purposes and also made available the pre-trained models (approximately 2.4 GB) and associated code. All of these are licensed under the Apache 2.0, a permissive license, in line with FAIR's dedication to open research.