Saturday, December 2, 2017

Install and set up s3fs in Ubuntu 16.04 to mount an s3 bucket

In this tutorial I am going to create a new s3 bucket then create a user with credentials for reading/writing/deleting documents in that bucket.  Then use that info to mount that bucket like a hard drive using s3fs on an Ubuntu 16.04 box. 

Create a bucket

Log into your console and go to the amazon s3 tool and click create bucket.

Give it a unique name and click next

For me I am going to leave the defaults here and click next

The default settings are good for me on this page as well.  I want my aws user to have read/write to this bucket, even though I am not going to use that user to mount the S3 bucket.   I also want to make the bucket private so “Do not grant public read access” .   Click Next

Click Create Bucket!

There is the bucket!

Create a user with credentials

Now I want to create a user with credentials who has permissions to read/write to this new bucket.

From the AWS console open up the IAM tool and click on users

Click add user.

Enter in a user name, select Programmatic Access and click Next Permissions.

Select Attach existing policies and search for s3.

Click on AmazonS3FullAccess and click on JSON

This JSON policy

    "Version": "2012-10-17",
    "Statement": [
            "Effect": "Allow",
            "Action": "s3:*",
            "Resource": "*"

Is close to what I want.  It gives read/write access, but to all my s3 buckets.  I want to tweak this slightly to limit it to a single bucket.

Here is a policy that will do what we want.

    "Version": "2012-10-17",
    "Statement": [
            "Effect": "Allow",
            "Action": "s3:ListBucket",
            "Resource": ["arn:aws:s3:::a-test-bucket-124568d"]
            "Effect": "Allow",
            "Action": [
            "Resource": ["arn:aws:s3:::a-test-bucket-124568d/*"]

Click on Create Policy

Click on JSON and paste the policy in.

Click Review Policy

Give it a name and click Create Policy

The policy has been created!

From the create user page select “Customer Managed” from the pull down menu.

Then click Refresh

Select the policy you just made and click Next: Review

Click Create User.

Record your Access Key ID and show your secret access key.

In my case

Secret:             Gcdpi2aETNuLsBxB1DKKU9g44qhpaAl6Eoviqreo

(don’t worry I am deleting this bucket and user after writing this how to)

Click Close you are done with this part.

Install s3fs

On Ubuntu 16.04 install the following tools

   > sudo apt-get install automake autotools-dev \
fuse g++ git libcurl4-gnutls-dev libfuse-dev \
libssl-dev libxml2-dev make pkg-config

Download the s3fs tool from github [1] and run the following to install it.

   > git clone
   > cd s3fs-fuse
   > ./
   > ./configure
   > make
   > sudo make install

Create the /etc/passwd-s3fs file

   > sudo vi /etc/passwd-s3fs

And place the ACCESS KEY ID : Secret Access KEY ID

So in my case I would put


Save it now chmod it to 640

   > sudo chmod 640 /etc/passwd-s3fs

Mount it!

Create a mount point and mount it!

   > sudo mkdir -p /s3/bucket-test
   > sudo s3fs -o allow_other a-test-bucket-124568d /s3/bucket-test

Its mounted now let me write a file to it.

Let me use /dev/urandom to create a 100MiB file with random data in it in a /tmp folder

   > cd /tmp
   > dd if=/dev/urandom of=random.txt count=1048576 bs=100

Now copy it over to the s3 bucket

   > cp /tmp/random.txt /s3/bucket-test/

Which fails because only root can write to it at the moment…
Let me use sudo

   > sudo cp /tmp/random.txt /s3/bucket-test/

OK that worked…

But how do I copy files over if I am not root?

I think I have an issue with /etc/fuse.conf file…

   > sudo vi /etc/fuse.conf

Uncomment out the user_allow_other line

Unmount the s3 drive and remount it.

   > sudo umount /s3/bucket-test
   > sudo s3fs -o allow_other a-test-bucket-124568d /s3/bucket-test

Now try and copy a file over.

   > cp /tmp/random.txt /s3/bucket-test/random2.txt




Mount using /etc/fstab

First let me unmount the s3 bucket

   > sudo umount /s3/bucket-test

Open and edit /etc/fstab

   > sudo vi /etc/fstab

And append the following line to the bottom of the file

s3fs#a-test-bucket-124568d /s3/bucket-test fuse retries=5,allow_other,url= 0 0

Here you can see the bucket and the mount point.

Now mount it

   > sudo mount /s3/bucket-test

Run a quick test

   > cp /tmp/random.txt /s3/bucket-test/random3.txt

That worked!

Also I do not think I need the user_allow_other in the fuse.conf file

   > sudo vi /etc/fuse.conf

Comment  out the user_allow_other line

I am just going to reboot at this point and see if it works with the new fuse.conf file and also if it automounts

   > sudo reboot now

Wahoo it is all working J

So that’s how you mount an S3 bucket as a drive in Ubuntu 16.04


[1]        s3fs github repo
                Accessed 11/2017


  1. This was very helpful to me. Thank you!
    Can you do a tutorial on how to do this on a CentOS based machine?
    I am running into some dependency issues trying to do this on a RHEL based EC2 instance.

    1. for CentOS

  2. for centOS

  3. when I tried to mount it returns with no error but mounting did not happen. can you advice what would be the issue? I am trying to mount it on ubuntu 18.02 instance

    1. I had the same issue, the _netdev mount option solved it. Perhaps this is worth a try ?

  4. This was an awesome how-to! Thank you!

  5. Thank you for this one ! I just connected my Wasabi S3 Storage. Works in the same way, just the URL is different.

    Example for an s3fs command:
    sudo s3fs bucket mountpoint -o passwd_file=/etc/passwd-s3fs -o url=
    -o allow_other

    It still have some issues (auto mount upon reboot does not work yet - no idea why) though - but will proceed with fingers crossed.

    1. ... and now even the mount problems are gone: the _netdev mount option solved the problem.

  6. This was super helpful to me! Thank you!!!

  7. great tutorial, worked perfectly.

  8. Great blog thanks for sharing Leaders in the branding business - Adhuntt Media is now creating a buzz among marketing circles in Chennai. Global standard content creation, SEO and Web Development are the pillars of our brand building tactics. Through smart strategies and customer analysis, we can find the perfect audience following for you right now through Facebook and Instagram marketing.
    digital marketing company in chennai
    seo service in chennai
    web designing company in chennai
    social media marketing company in chennai

  9. If only it were that simple: archaeology involves collaboration and working with teams, and every excavation is a project, so management skills are critical to success. nice page
    Ai & Artificial Intelligence Course in Chennai
    PHP Training in Chennai
    Ethical Hacking Course in Chennai Blue Prism Training in Chennai
    UiPath Training in Chennai

  10. Thanks for your extraordinary blog. Your idea for this was so brilliant. This would provide people with an excellent tally resource from someone who has experienced such issues.

    AWS training in Chennai

    AWS Online Training in Chennai

    AWS training in Bangalore

    AWS training in Hyderabad

    AWS training in Coimbatore

    AWS training

    AWS online training

  11. thank you for sharing we are looking forward for more
    hadoop Training in Hyderabad

  12. you have executed an uproarious undertaking upon this text. Its completely change and very subjective. you have even figured out how to make it discernible and simple to make a get accord of into. you have a couple of definite composing dexterity. much appreciated likewise a lot. Movavi Video Editor Serial Key

  13. I go to your blog every now and again and counsel it to the total of people who wanted to highlight on happening their comprehension resulting to ease. The way of composing is perfect and in addition to the substance material is highest point score. gratitude for that insight you give the perusers! Stellar Data Recovery Crack


  14. I was hundreds of miles away on your special day. I hope you enjoyed your day with a big cake. What To Say In A Belated Birthday Card