Mahi is an all-in-one HTTP service for file uploading, processing, serving, and storage. Mahi supports chunked, resumable, and concurrent uploads. Mahi uses Libvips behind the scenes making it extremely fast and memory efficient.
Mahi currently supports any s3 compatible storage, which includes (AWS s3, DO Spaces, Wasabi, Backblaze B2). The specific storage engine can be set when creating an application.
Mahi supports different databases for storing file meta-data and analytics. Currently, the 2 supported databased are PostgreSQL and BoltDB. The database of choice can be provided via the config file.
- Uploads upload files easily via our Web API.
- Chunked, Concurrent, Resumable Uploads upload large files by chunking them and uploading them to our Web API.
- Flexible File Storage store your files in S3, Spaces, Wasabi, B2 with more options coming soon.
- Image Processing resize, convert, and crop with ease.
- Usage Analytics get insights on how many transformations, how much storage and bandwidth you are processing.
Libvips must be installed on your machine.
sudo apt install libvips libvips-dev libvips-tools
brew install vips
For other systems check out instructions here.
Installing mahid server.
go get -u github.com/threeaccents/mahi/...
This will install the
mahid command in your
If no config is passed Mahi will look for a
mahi.toml file in the current directory.
Mahi has the concept of applications. Each application houses specific files and the storage engine for those files. This makes Mahi extremely flexible to use for different projects. If on one project you decide to use s3 as your storage engine and another DO Spaces, Mahi easily handles it for you.
Applications can be created via our Web API.
Files are uploaded to Mahi via
multipart/form-data requests. Along with passing in the file data, you must also provide the
application_id. Mahi will handle processing and storing the file blob in the application's storage engine along with storing the file meta-data in the database. To view an example upload response check out the Web API
Large File Uploads
When dealing with large files, it is best to split the file into small chunks and upload each chunk separately. Mahi easily handles chunked uploads storing each chunk and then re-building the whole file. Once the whole file is re-built Mahi uploads the file to the application's storage engine. To view an example upload response check out the Web API
Other benefits of chunking up files are the ability to resume uploads and uploading multiple chunks concurrently. Mahi handles both scenarios for you with ease.
File Transformations (More Coming Soon)
Mahi supports file transformations via URL query params. Currently, the supported operations are:
- Resize (width, height)
- Smart Crop
- Black and White
- Quality(JPEG), Compression(PNG)
- Format conversion
format is based on the file extension. To transform a png to webp, just use the .webp extension.
All queries can be used together. For example, to resize the width, make the image black and white, and change the format to webp the params would look like this:
Mahi currently tracks these stats for both specific applications and the service as a hole:
- Transformations: Total transformations
- Unique Transformations: Unique transformations per file.
- Bandwidth: Bytes served.
- Storage: Bytes stored.
- File Count: Total files.
These stats can be retrieved via our Web API.
Mahi's is configured via a toml file. Here are toml config examples. Configuration options include:
- db_engine:string(default: bolt) The main database for mahi. Valid options are
bolt. This is not to be confused with the storage engine. Storage engine is set per application via the Web API.
- port:int(default: 4200) the port to run mahi on.
- https:boolean(default: false) configures server to accept https requests.
- ssl_cert_path:string path to ssl certificate. Only required if
httpsis set to true.
- ssl_key_path:string path to ssl key. Only required if
httpsis set to true.
- auth_token:string token for authenticating requests
- aes_key:string key for use with AES-256 encryption. This is used to encrypt storage secrets.
- chunk_upload_dir:string(default: ./data/chunks) directory for storing chunks while an upload is happening. Once an upload is completed, the chunks are deleted.
- full_file_dir:string(default: ./data/files) full_files are temp files used while building chunks or downloading files from the storage engine. These temp files are removed once the request is completed.
- max_chunk_size:int64(default: 10MB) max size of a file chunk in bytes.
- max_file_size_upload:int64(default: 50MB) max size of a file for a regular upload in bytes.
- max_transform_file_size:int64(default: 50MB) max size of a file that can be transformed in bytes.
- bolt(only used if
db_engineis set to bolt
- dir:string(default: ./data/mahi/mahi.db) directory for bolt db file.
- postgresql(only used if
db_engineis set to postgres)
- database:string(default: mahi) name of database.
- host:string(default: localhost) host of database.
- port:int(default: 5432) port of database.
- user:string(default: mahi) username of database.
- password:string(default: ) password of database.
- max_conns:int(default: 10 connections per CPU) maximum connections for database pool.
To use Postgres the necessary data tables must be created. SQL files are located in the migrations folder. In the future, Mahi will come with a migrate command that will automatically handle creating the necessary tables for you. For now, you have 2 options. Install tern,
cd into the migrations folder, and run
tern migrate. The second option is just to copy and paste the SQL provided directly in a GUI or command line instance of Postgres.