Using DigitalOcean Spaces or Amazon S3 for file storage in XF 2.1+

Using DigitalOcean Spaces or Amazon S3 for file storage in XF 2.1+ V2.3.0

No permission to download
Author jessy
Creation date
Headline
The download is only compatible with XenForo 2.1 and above.
Short Description
FREE Download Premium and Nulled Using DigitalOcean Spaces or Amazon S3 for file storage in XF 2.1+ V2.3.0 version from NullPro Community. It is zip Extention type and 627.3 KB File size. From Using DigitalOcean Spaces or Amazon S3 for file storage in XF 2.1+ have 2 Updates, 284 Views.
XenForo Version Compatibility

The download is only compatible with XenForo 2.1 and above.

Why this guide?

Since XenForo 2.0.0 we have supported remote file storage using an abstracted file system named Flysystem. It's called an abstracted file system as it adds a layer of abstraction between the code and a file system. It means that it provides a consistent API for performing file system operations so that whether the file system is a local disk-based file system or a distributed and remotely accessible file system, our code calls the same functions and Flysystem takes care of the rest.

As useful as that is, it isn't the most obvious or straightforward thing to set up so this guide and accompanying add-on will help.

So, if you're planning to make use of the video uploads function in XF 2.1 or above and you're worried about increased disk space requirements this will help.


Making the required files available

Although it is possible for you to download the files and set up things like the autoloader yourself, you will probably prefer to simply download the add-on that is attached to this resource. You can install the add-on in the usual fashion.


Before you start

If you're setting this up on an existing site, you will need to manually move your existing files over. There's a section about that at the end. While you are moving the existing files, setting things up and testing, we recommend closing the forum first.


Setting up DigitalOcean Spaces

We'll cover this first as it is the most straightforward to set up. If you'd prefer to use Amazon S3 skip ahead to the Setting up Amazon S3 section below.
  1. Go to the DigitalOcean Cloud page and sign up or log in.
  2. At this point, if you're new to DigitalOcean, you may need to set up billing.
  3. You will now be able to create a new project.
  4. Click the "Start using Spaces" link.
  5. Choose your datacenter region (I have chosen Amsterdam).
  6. Leave "Restrict File Listing" selected.
  7. Choose a unique name (I have chosen "xftest")
  8. Click "Create a space"
Now the space is created, you should have an endpoint URL, similar to: https://xftest.ams3.digitaloceanspaces.com. Note this down for later.

Now we need to create some API credentials. To do this:
  1. Click "Manage" in the left sidebar.
  2. Click "API".
  3. In the "Spaces access keys" section click "Generate New Key".
  4. Type a name for the key (Again, I have chosen "xftest") and save.
This will give you a key and a secret. Note them down.

Configuring XF to use DigitalOcean Spaces

We now need to configure XF to use DigitalOcean Spaces for file storage. We'll start with what usually goes into the data directory first. This generally includes attachment thumbnails and avatars.

Open your src/config.php file.

First we need to configure the Amazon S3 client (the DigitalOcean Spaces API is compatible with the Amazon AWS SDK).

We will do this using a closure so that we can reuse the same code and we only have to type it out once:
PHP:
$s3 = function()
{
return new \Aws\S3\S3Client([
'credentials' => [
'key' => 'ABC',
'secret' => '123'
],
'region' => 'ams3',
'version' => 'latest',
'endpoint' => 'https://ams3.digitaloceanspaces.com'
]);
};
Note that the key and secret are what you noted down after setting up the "Spaces access key" earlier. The region can be inferred from the endpoint URL you noted down earlier. It's the part after the first . in the URL, in my case it is ams3. The endpoint is the same endpoint URL minus the unique name you chose.

Next we need to set up the actual Flysystem adapter to use the S3 client:
PHP:
$config['fsAdapters']['data'] = function() use($s3)
{
return new \League\Flysystem\AwsS3v3\AwsS3Adapter($s3(), 'xftest', 'data');
};
Finally, we need to ensure that avatar and attachment thumbnail URLs are prepended with the correct URL. This requires the endpoint URL you noted down earlier, again:
PHP:
$config['externalDataUrl'] = function($externalPath, $canonical)
{
return 'https://xftest.ams3.digitaloceanspaces.com/data/' . $externalPath;
};
At this point, everything should be working in terms of new uploads. Don't be alarmed if you notice that avatars and thumbnails are missing; if you have existing files, they will need to be moved over manually which we'll go through later.

First, we need to test that the configuration works. Simply go and upload a new avatar. The avatar will now be stored and served remotely!

If you check your DigitalOcean Spaces account now, you should see that new folders have been created containing your new avatar:

187981


Success! ? But we're only half way there!

We now need to add support for the internal_data directory stuff too. Generally, this is attachments and any other stuff that should be "private". Back to config.php and the code to add is very similar:
PHP:
$config['fsAdapters']['internal-data'] = function() use($s3)
{
return new \League\Flysystem\AwsS3v3\AwsS3Adapter($s3(), 'xftest', 'internal_data');
};
Now try to upload an attachment to a post and, much like before, you should now see additional files and folders in your Spaces file browser. ?


Setting up Amazon S3
  1. Go to the AWS Management Console page and sign up or log in.
  2. In the "AWS services" section type "S3" to go to the "S3 Console".
  3. Click "Create bucket".
  4. Choose a bucket name (I have chosen xftest).
  5. Choose a region (I have chosen EU London).
  6. Accept any further default options until the bucket is created.
  7. You now need to go to the "IAM" console.
  8. Click "Add user".
  9. Pick a username (yep, I used xftest again ?).
  10. Set the access type to "Programmatic".
  11. To set permissions, click the "Attach existing policies directly" tab followed by the "Create policy" button.
  12. IAM and the various policies and permissions can be fairly daunting. We can make it a bit easier, though you may have different requirements. On this page there is a tab called "JSON". Paste the following in there, replacing YOUR-BUCKET-NAME with the bucket name you chose earlier:
JSON:
{
"Version": "2012-10-17",
"Statement": [
{
"Effect": "Allow",
"Action": [
"s3:ListBucket",
"s3:GetObject",
"s3:GetObjectAcl",
"s3:putObject",
"s3:putObjectAcl",
"s3:ReplicateObject",
"s3:DeleteObject"
],
"Resource": [
"arn:aws:s3:::YOUR-BUCKET-NAME",
"arn:aws:s3:::YOUR-BUCKET-NAME/*"
]
}
]
}
  1. Click "Review policy" give it a name and save.
  2. Go back to the previous "Add user" page, click the "Refresh" button and search for the policy you just created.
  3. Click "Next", followed by "Create user".
This will give you a key and a secret. Note them down.


Configuring XF to use Amazon S3

We now need to configure XF to use Amazon S3 for file storage. We'll start with what usually goes into the data directory first. This generally includes attachment thumbnails and avatars.

Open your src/config.php file.

We will do this using a closure so that we can reuse the same code and we only have to type it out once:
PHP:
$s3 = function()
{
return new \Aws\S3\S3Client([
'credentials' => [
'key' => 'ABC',
'secret' => '123'
],
'region' => 'eu-west-2',
'version' => 'latest',
'endpoint' => 'https://s3.eu-west-2.amazonaws.com'
]);
};
Note that the key and secret are what you noted down after setting up the IAM user earlier. The region can be inferred from the S3 endpoint URL.

Next we need to set up the actual Flysystem adapter to use the S3 client:
PHP:
$config['fsAdapters']['data'] = function() use($s3)
{
return new \League\Flysystem\AwsS3v3\AwsS3Adapter($s3(), 'xftest', 'data');
};
Finally, we need to ensure that avatar and attachment thumbnail URLs are prepended with the correct URL:
PHP:
$config['externalDataUrl'] = function($externalPath, $canonical)
{
return 'https://xftest.s3.eu-west-2.amazonaws.com/data/' . $externalPath;
};
At this point, everything should be working in terms of new uploads. Don't be alarmed if you notice that avatars and thumbnails are missing; if you have existing files, they will need to be moved over manually which we'll go through later.

First, we need to test that the configuration works. Simply go and upload a new avatar. The avatar will now be stored and served remotely!

If you check your bucket file browser now, you should see that new folders have been created containing your new avatar:

187995


Success! ? But we're only half way there!

We now need to add support for the internal_data directory stuff too. Generally, this is attachments and any other stuff that should be "private". Back to config.php and the code to add is very similar:
PHP:
$config['fsAdapters']['internal-data'] = function() use($s3)
{
return new \League\Flysystem\AwsS3v3\AwsS3Adapter($s3(), 'xftest', 'internal_data');
};
Now try to upload an attachment to a post and, much like before, you should now see additional files and folders in your bucket file browser. ?


Moving existing files to DigitalOcean Spaces or Amazon S3

So, you now have remotely hosted files. At least, you do from this point onwards. But what about all of the existing files you have?

Thankfully there are several ways to interact with Spaces and S3 in order to make moving your existing content over very easily. Although this is a one-time operation, depending on the number and size of the files, it could take a significant amount of time.

There are a few ways to manage this process, but arguably the best approach is to use a tool by the name of s3cmd which is a popular cross-platform command-line tool for managing S3 and S3-compatible object stores.

It should be possible whether you are using Spaces or S3 to install the s3cmd tool on your server and run the commands to copy the files across to their new home.

Rather than rehashing something that has already been written, I'll leave you with the following guide from DigitalOcean which goes through how to migrate your existing files using s3cmd.

s3cmd 2.x Setup :: DigitalOcean Product Documentation

s3cmd is a popular cross-platform command-line tool for managing S3 and S3-compatible object stores. To use s3cmd, you will need: s3cmd version 2.0.0+ or higher. You can check your version with s3cmd --version. Versions from package managers may be out of date, so we recommend using the s3cmd...
www.digitalocean.com
Note: When copying your existing data files across, they will need to be made public. You can do this by setting the ACL to public while copying:

Examples of s3cmd 2.x Usage with DigitalOcean Spaces :: DigitalOcean Documentation

Examples of s3cmd commands for common actions with Spaces, like uploading and managing files, setting permissions, and encryption.
www.digitalocean.com
Author
jessy
Views
284
Extention type
zip
File size
627.3 KB
First release
Last update
Rating
0.00 star(s) 0 ratings
Link was Broken? Please  Send Message to NP Team with Ticket, You will get it very quickly!
Support Developer If you are satisfied with your test or project have earn money successfully, Maybe you can click more information button to support with buying.
More resources from jessy

Latest updates

V2.3.0 - V2.3.0
This update is a very minor maintenance release. As well as including the latest version of the...
2.2.1 - 2.2.1
Updates Amazon AWS SDK to version 3.209.16 The last update to this add-on updated the Amazon...

Similar resources

Upload & Host Images on your forum using Add.Pics for free N
Remote Image upload & storage using add.pics for free image hosting.
0.00 star(s) 0 ratings
Views
407
Updated
Back
Top