Automate Midjourney with Make.com: What’s the Exact Blueprint?

Automate Midjourney with Make.com and ImagineAPI.dev. This blueprint uses ChatGPT Vision to easily auto-generate, auto-name, and save images to Google Drive.

Spread the love
Bikers rights gif via giphy
"bikers rights" portlandia gif via giphy

TL;DR

You need two separate Make.com scenarios for this to work: one to trigger Midjourney generation, and one to catch completed images via webhook, process them, and save them.

ImagineAPI.dev is the Midjourney API connector that works without weird crypto payment hoops (looking at you, UserAPI.ai).

ChatGPT Vision can analyze each generated image and create a descriptive, sanitized filename automatically. No more “image_001.png” nonsense.

The entire workflow fits under Make.com’s free tier (under 1,000 operations/month) and costs roughly $27–$47/month total between Midjourney and ImagineAPI depending on your plans.

I’m gonna show you exactly how I use Midjourney within Make.com to make images while I’m sleeping. 🤙 And this is proof, not some theoretical “you could do this” situation. This is a real workflow I built, tested, broke, fixed, and now run on autopilot.

If you’ve been manually generating Midjourney images one by one, saving them to your desktop, and renaming them something like “final_final_v3_REAL.png”, then honestly we need to talk. 😤

The whole thing runs on two Make.com scenarios, a Midjourney API bridge called ImagineAPI.dev, and a clever little ChatGPT Vision trick that automatically names your images based on what’s actually in them. And I’m giving you the full blueprint, completely for free.

If you’re already building automations with Make.com, this is going to slot right into your existing setup.

https://www.youtube.com/watch?v=

Create images even while you sleep using make.com and Midjourney.

Why You Need Two Scenarios (Not One)

This is the thing that trips people up first. You can’t just build one long Make.com scenario for this. It has to be split into two.

Thumbnail showing automation workflow for midjourney and make. Com
Create images even while you sleep using make. Com and midjourney.

Why? Because Midjourney doesn’t generate images instantly. When you send a prompt, it goes through that whole percentage-loading thing you know from Discord, 25%, 30%, 40%, all the way to 100%. Your first scenario fires off the prompt and then it’s done; it’s not sitting around waiting.

The second scenario listens for a webhook that fires only when the image is actually complete.

This is broken up into two scenarios, and it has to be broken up into two scenarios for this to work. One triggers. One catches. That’s the architecture, and I can’t stress that enough.

Warning callout icon.

Warning

Don’t try to combine these into a single scenario. The asynchronous nature of Midjourney’s image generation means your first scenario will time out waiting for results. The webhook-based second scenario is the only reliable way to catch completed images.

The first scenario is set to run on demand (because I’m just showing the Midjourney piece here; in production you’d have upstream modules generating dynamic prompts). The second scenario must be set to immediately so it catches that webhook the moment the image finishes generating.

This has to be immediate and the first one has to be on demand. Get that backwards and nothing works.

What You Actually Need (And What It Costs)

Before we get into the build, here’s your shopping list:

ToolCostPurpose
Make.com accountFree tier worksAutomation platform that runs both scenarios
Midjourney subscriptionStarting at $10/month (Basic)The AI image generator itself
ImagineAPI.devStarting at $16/monthThe API bridge connecting Make.com to Midjourney
OpenAI API access~$1/month or lessChatGPT Vision for image analysis + filename generation
Google DriveFreeWhere your finished images land
Tools, costs, and what each part does in the workflow.

So you’re looking at roughly $27–$47/month depending on your chosen plans to run this entire system. I have the $30 Midjourney plan because I use this quite a bit, but the $10 Basic plan works fine to get started.

Check ImagineAPI.dev’s current pricing for the latest rates, as plans may have changed.

Why ImagineAPI.dev and Not UserAPI.ai

I have opinions here, and I’m not gonna sugarcoat them.

A lot of tutorials out there point you toward UserAPI.ai for connecting Midjourney to Make.com. And look, I tried it. I genuinely did. There are a lot of videos showing you how to use it, but the problem is that once I got to the pricing, it tells me how much I have to pay, but then it needs me to pay with crypto or something like that.

And I was just like… no.

I’m not trying to join a new payment processor to buy crypto so that I could use it for this platform. I’m just not trying to do that. That really turned me off from using UserAPI.ai, so I didn’t go with them.

ImagineAPI.dev is straightforward. Credit card, monthly plan, done. They also have a native Make.com module, which is super handy. You’re not messing around with raw HTTP requests and custom headers.

You just drop in the module, add your API key, and you’re rolling. I’ve gone back and forth with the developer on support issues and it’s been genuinely great.

ImagineAPI.dev vs. UserAPI.ai

ImagineAPI.dev ✅

  • Native Make.com module
  • Standard credit card payment
  • Responsive developer support
  • Simple API key setup

UserAPI.ai ❌

  • Requires custom HTTP configuration
  • Crypto payment option can be a barrier
  • More tutorials available, but messier experience
  • Additional setup complexity

Worth knowing: there’s no official Midjourney API. Midjourney has never released a fully public, generally available API. Every “Midjourney API” you see is an unofficial third-party service that’s figured out how to interface with Midjourney’s backend.

ImagineAPI.dev is upfront about this, calling itself “The (unofficial) Midjourney API.”

Building Scenario 1: The Trigger

This is the simple one. You’re setting up a scenario that sends a prompt to Midjourney via ImagineAPI.dev.

Step 1: Add the ImagineAPI module to your scenario. Search for it in Make.com’s module library. It’s called “Generate an Image.” Click it, add it to your scenario, done.

Step 2: Create a connection. Go to your ImagineAPI.dev dashboard, find your API keys section, copy your key, and paste it into the connection setup in Make.com. That’s quite literally it for authentication.

Step 3: Add your prompt. In my demo I just typed something like “image of a dog with wings” as plain text. But in a real production workflow, you’d have upstream modules—maybe a ChatGPT module generating prompts, or a Google Sheets row feeding in data—piping a dynamic variable into this field.

You’re using Make.com, so you want dynamic variables. You don’t want to come here and manually type this out every time.

Here’s a little trick I do: I always have some pre-loaded style modifiers appended to my prompts. Things like “HD detail, duotone illustration, comic book style” and specific colors that I really want to emphasize.

You can also add your regular Midjourney parameters (like --ar 16:9 or --v 6.1) directly in the prompt field. They work exactly the same as they would in Discord.

Once you hit run, this scenario fires the prompt off to ImagineAPI, which sends it to Midjourney, and… that’s it. Scenario 1 is done. It just sent data and moved on with its life. It looks like nothing happened, right? It just ran. But what it actually did is launch the generation on Midjourney’s side.

Building Scenario 2: The Webhook Catcher

This is where the real magic happens, and it’s got more moving parts.

Setting Up the Webhook

First, you need a custom webhook in Make.com. Create a new scenario, add a Webhook module (custom webhook), click save, and it’ll generate a URL. Copy that URL.

Now go to your ImagineAPI.dev backend, navigate to Webhooks, click “Add a Webhook,” and paste that Make.com URL in there. What this does is tell ImagineAPI: “When an image is done, send the data here.”

Info icon.

Did You Know?

There’s a filter between the webhook module and the next step. Set it to check that payload status = completed because ImagineAPI sends updates at multiple stages (25%, 40%, etc.). You only want to proceed when the image is actually finished.

Getting the Image Data

Once the webhook confirms the image is complete, the next module grabs the image data using the ID from the payload. Midjourney generates four image variations per prompt (the classic 2×2 grid), so you’re getting four separate upscaled URLs back.

To process all four, you need an iterator module. The iterator takes the array of upscaled URLs and splits them into individual bundles, one per image. And here’s a specific thing that caught me up: when you’re configuring the iterator, select the plain upscaled URLs array, not the nested version.

Get it wrong and the whole thing chokes.

The HTTP Get

After the iterator, an HTTP “Get a File” module downloads each image using the URL value from the iterator. This pulls the actual image file into Make.com’s memory so you can do stuff with it. The variable here is URL, and the URL is the value of the iterator. Straightforward.

The ChatGPT Vision Trick (This Is the Cool Part)

This part is totally optional, but it’s my favorite piece of the whole build.

After downloading each image, I pass it to OpenAI’s “Analyze Image with Vision” module. The prompt is dead simple: “What is this image?” ChatGPT Vision looks at the image and comes back with something like: “The image appears to be an aesthetic representation of a dog with wings, possibly styled in a pop art format.”

Each of the four images gets its own unique description because they’re all different images. The Vision model is actually looking at what’s in each one, not just regurgitating your original prompt.

Error icon.

Critical Error

Make sure you select “Analyze Image with Vision.” It’s a specific OpenAI module, different from the regular text completion modules. The standard text-only modules can’t process images.

Then I pass that description to a second OpenAI module, this time just a regular GPT-4o-mini completion, with a prompt like: “Based on that description, create a sanitized file name that’s less than 100 characters.”

I use GPT-4o-mini here because from what I saw it’s the cheapest model. I can run this for a month and pay less than a dollar.

Saving to Google Drive

The last module takes each image file and uploads it to a specified Google Drive folder, using the ChatGPT-generated filename as the actual file name.

So instead of downloading four images named image_001.png through image_004.png, I end up with files like “dog-with-wings-pop-art-style.png” and “winged-dog-duotone-illustration.png” sitting in my Drive—organized and descriptively named—without me touching a thing. It’s great.

The whole workflow, from prompt to four named, organized images in Google Drive, runs automatically.

The Operations Math

One thing people worry about with Make.com is operations. The free tier gives you 1,000 per month, and people panic about burning through them.

Here’s how this workflow breaks down per run: you’ve got the trigger (1 operation), the webhook catch (1), getting image data (1), the iterator processing four images (4), four HTTP gets (4), four Vision analyses (4), four filename generations (4), and four Google Drive uploads (4). That’s roughly 23 operations per run.

At 1,000 free operations, you can run this about 43 times per month before you’d need to upgrade. That’s around 172 images. For most people doing content creation, that’s more than enough to stay on the free tier.

Do you ever wonder if you could create stunning images while you sleep, saving time and effort, or is it just me? 😴
Do you ever wonder if you could create stunning images while you sleep, saving time and effort, or is it just me? 😴

Frequently Asked Questions

Yes, as long as your Midjourney subscription covers commercial use, which all paid Midjourney plans do (the free trial does not grant commercial rights). The automation layer doesn’t change the licensing terms of the images Midjourney generates. Check Midjourney’s terms of service for the latest details.

ImagineAPI will send a webhook with a status other than “completed”. Because of the filter between the webhook and the next module, the scenario simply won’t proceed. It won’t error out or charge you extra operations. You’d just need to re-trigger the first scenario.

Absolutely. Make.com has modules for Dropbox, OneDrive, Airtable, Notion, and dozens of other storage platforms. Replace the final Google Drive module with whatever you prefer, and the rest of the workflow stays identical.

No. Everything here is done through Make.com’s visual drag-and-drop interface. The closest thing to “code” you’ll encounter is pasting in your ImagineAPI key and writing ChatGPT prompts. If you’ve ever built a Make.com scenario before, you already have the skills for this.

Yes—this is ideal for production. Add a Google Sheets or Airtable module before the ImagineAPI trigger, and it’ll pull prompts from rows automatically. You could queue up 50 prompts in a spreadsheet and let the whole thing run through them while you sleep.

The whole point of this build is getting Midjourney out of the “manually click around in Discord” phase and into something that works for you in the background. Two scenarios, one API bridge, and a ChatGPT trick for filenames. That’s the entire system.

It’s not complicated once you see it laid out, and honestly the hardest part was finding an API provider that didn’t make me want to pull my hair out (sorry, UserAPI.ai).

If you’re a content creator, marketer, or solo entrepreneur who needs consistent image output without babysitting Discord all day, this is the workflow. Set up the two scenarios, connect your ImagineAPI account, add the Vision module for those auto-generated filenames, and point it all at a Google Drive folder.

Then go do literally anything else. The images will be there when you get back.

Leave a Comment

Frequently asked questions (FAQ)

LiaisonLabs is your local partner for SEO & digital marketing services in Mount Vernon, Washington. Here are some answers to the most frequently asked questions about our SEO services.

SEO (Search Engine Optimization) is the process of improving your website's visibility in search engines like Google. When potential customers in Mount Vernon or Skagit County search for your products or services, SEO helps your business appear at the top of search results. This drives more qualified traffic to your website—people who are actively looking for what you offer. For local businesses, effective SEO means more phone calls, more foot traffic, and more revenue without paying for every click like traditional advertising.

A local SEO partner understands the unique market dynamics of Skagit Valley and the Pacific Northwest. We know the seasonal patterns that affect local businesses, from tulip festival tourism to agricultural cycles. Local expertise means we understand which keywords your neighbors are searching, which directories matter for your industry, and how to position your business against local competitors. Plus, we're available for in-person meetings and truly invested in the success of our Mount Vernon business community.

SEO is a long-term investment, and most businesses begin seeing meaningful results within 3 to 6 months. Some quick wins—like optimizing your Google Business Profile or fixing technical issues—can show improvements within weeks. However, building sustainable rankings that drive consistent traffic takes time. The good news? Unlike paid advertising that stops the moment you stop paying, SEO results compound over time. The work we do today continues delivering value for months and years to come.

SEO pricing varies based on your goals, competition, and current website health. Local SEO packages for small businesses typically range from $500 to $2,500 per month, while more comprehensive campaigns for competitive industries may require a larger investment. We offer customized proposals based on a thorough audit of your website and competitive landscape. During your free consultation, we'll discuss your budget and create a strategy that delivers measurable ROI—because effective SEO should pay for itself through increased revenue.

Both aim to improve search visibility, but the focus differs significantly. Local SEO targets customers in a specific geographic area—like Mount Vernon, Burlington, Anacortes, or greater Skagit County. It emphasizes Google Business Profile optimization, local citations, reviews, and location-based keywords. Traditional SEO focuses on broader, often national rankings and prioritizes content marketing, backlink building, and technical optimization. Most Mount Vernon businesses benefit from a local-first strategy, though many of our clients combine both approaches to capture customers at every stage of their search journey.

Absolutely! SEO and paid advertising work best as complementary strategies. Google Ads deliver immediate visibility and are great for testing keywords and driving quick traffic. SEO builds sustainable, long-term visibility that doesn't require ongoing ad spend. Together, they create a powerful combination—ads capture immediate demand while SEO builds your organic presence over time. Many of our Mount Vernon clients find that strong SEO actually improves their ad performance by increasing Quality Scores and reducing cost-per-click, ultimately lowering their total marketing costs while increasing results.