If you're looking for an API, you can choose from your desired programming language.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
import requests
api_key = "YOUR_API_KEY"
url = "https://api.segmind.com/v1/minimax-ai-director"
# Prepare data and files
data = {}
files = {}
data['prompt'] = "A bear playfully poses for a selfie with a man, who smiles and says, ‘Hi!’ as he looks into the camera, [Pedestal up] sharing a lighthearted moment with his friend.[Zoom out]"
# For parameter "first_frame_image", you can send a raw file or a URI:
# files['first_frame_image'] = open('IMAGE_PATH', 'rb') # To send a file
# data['first_frame_image'] = 'IMAGE_URI' # To send a URI
data['prompt_optimizer'] = True
headers = {'x-api-key': api_key}
response = requests.post(url, data=data, files=files, headers=headers)
print(response.content) # The response is the generated image
Text prompt for video generation. Camera movement instructions can be added using square brackets (e.g. [Pan left] or [Zoom in]). You can use up to 3 combined movements per prompt. Supported movements: Truck left/right, Pan left/right, Push in/Pull out, Pedestal up/down, Tilt up/down, Zoom in/out, Shake, Tracking shot, Static shot. For example: [Truck left, Pan right, Zoom in]
First frame image for video generation
Use prompt optimizer
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Generate high-definition videos with Minimax video-01-director using specific camera movements. The minimax/video-01-director model supports 720p resolution at 25fps and creates videos up to 5 seconds. Control camera movements using bracketed commands or natural language within the promp
AI Video Generation: Creates videos from text or images.
Cinematic Camera Movements: Allows control over camera movements like pan, tilt, zoom, and tracking.
High Definition: Supports 720p resolution at 25fps.
Camera Movement Control:
Use bracketed commands (e.g., [Pan left, Tilt up]).
Incorporate natural language descriptions within the prompt.
Supported Camera Movements: Includes horizontal, vertical, depth, and special movements.
Horizontal: [Truck left], [Truck right], [Pan left], [Pan right].
Vertical: [Pedestal up], [Pedestal down], [Tilt up], [Tilt down].
Depth: [Push in], [Pull out], [Zoom in], [Zoom out].
Special: [Tracking shot], [Shake], [Static shot].
Preset Shot Combinations: Offers preset combinations for circling, walking, and dynamic shots.
Circling shots: [truck left, pan right, tracking shot] (left circling), [truck right, pan left, tracking shot] (right circling).
Walking shots: [truck left, tracking shot] (left walking), [truck right, tracking shot] (right walking).
Dynamic shots: [push in, pedestal up] (upward tilt), [truck left, pedestal up] (scenic shot), [pan right, zoom in] / [pan left, zoom in] (stage shots), [pedestal down, tilt up] (downward tilt)
Be specific in camera movement descriptions.
Place movement commands at logical points in the prompt.
Combine natural language with bracketed commands for precise control.
Limit combined movements to 3 or fewer.
Start with a clear subject description.
Include camera movements in brackets where they should occur.
Add details about style, mood, or environment.
Consider sequential timing of movements
SDXL ControlNet gives unprecedented control over text-to-image generation. SDXL ControlNet models Introduces the concept of conditioning inputs, which provide additional information to guide the image generation process
Story Diffusion turns your written narratives into stunning image sequences.
The SDXL model is the official upgrade to the v1.5 model. The model is released as open-source software
CodeFormer is a robust face restoration algorithm for old photos or AI-generated faces.