Compare commits
105 Commits
MailcowBac
...
main
Author | SHA1 | Date | |
---|---|---|---|
46b29dfce9 | |||
47adaceed0 | |||
475a0d0dd2 | |||
818483f92e | |||
c8b328e707 | |||
fb4dcab29a | |||
5382dd26d9 | |||
bd36035a9f | |||
389f5ef059 | |||
8d10a1a3b9 | |||
1ac614ee60 | |||
d0141e3296 | |||
17bd522ff7 | |||
5aa9068556 | |||
|
aa2c17799c | ||
|
1334975cab | ||
|
19d1eb1587 | ||
|
4aae61f97e | ||
|
5ea25529b9 | ||
|
de76918e81 | ||
|
16f4c11f91 | ||
|
0bc0f8e5f9 | ||
|
21e4a6774c | ||
|
7e739d96a1 | ||
|
61e505952d | ||
|
1939962b75 | ||
|
f2231c7074 | ||
|
2747b10040 | ||
|
876f8eee9e | ||
|
3a3b02685b | ||
|
85e34c73a3 | ||
|
9e592633b9 | ||
|
e127616800 | ||
|
053e45affc | ||
|
fb58c06489 | ||
|
010ab1dceb | ||
|
f28ad5e6e7 | ||
|
06df565348 | ||
|
ea0d98734a | ||
|
b07ad2c0ee | ||
|
84605567cd | ||
|
8c6019f7da | ||
|
363c970f2f | ||
|
cc30581382 | ||
|
c9b85d084e | ||
|
bf18c80f94 | ||
|
764f35a695 | ||
|
ba212f738f | ||
|
59cc9a5d66 | ||
|
7b3ff7f6ae | ||
|
78e1ed2539 | ||
|
af4fcf3cf3 | ||
|
561437d667 | ||
|
e4cd29fd42 | ||
|
54ca63657f | ||
|
de96d55268 | ||
|
1d793dc291 | ||
|
7ce419d3eb | ||
|
eadba0bee9 | ||
|
9a363d8510 | ||
|
5c7064e2bd | ||
|
1cbbf3d66d | ||
|
2b5f8a0887 | ||
|
85fae7f6c8 | ||
|
32daba03b3 | ||
|
a1aff03e2d | ||
|
ee254a4356 | ||
|
0680485ab8 | ||
|
f89d8f26a7 | ||
|
f90b97fc9b | ||
|
8680f91572 | ||
|
b6c05e4ccb | ||
|
7293a29a05 | ||
|
8486c4b7d2 | ||
|
f57841b69a | ||
|
4d407887d5 | ||
|
6f59804cfc | ||
|
b5247a085e | ||
|
2432139f3d | ||
|
5f0eddf22a | ||
|
09c1d903ba | ||
|
51928d4327 | ||
|
a7f506efb0 | ||
|
0993083a7b | ||
|
5448848c36 | ||
|
5667204378 | ||
|
ca1186d855 | ||
|
f8561a3aab | ||
|
1b68616e54 | ||
|
51120491d5 | ||
|
abdbb44922 | ||
|
ca62d81368 | ||
|
72e396fa64 | ||
|
992f8ef16e | ||
|
9d7b7e1a94 | ||
|
4fb6315643 | ||
|
d9c662bb45 | ||
|
e9884a347f | ||
|
8a5a87db12 | ||
|
0d32aaf40a | ||
|
bb952a3aba | ||
|
dccf5e17d0 | ||
|
35ce452dec | ||
|
604d3da07d | ||
|
a5bc4549e9 |
53
.gitattributes
vendored
53
.gitattributes
vendored
@ -10,17 +10,17 @@
|
|||||||
|
|
||||||
# Documents
|
# Documents
|
||||||
*.bibtex text diff=bibtex
|
*.bibtex text diff=bibtex
|
||||||
*.doc diff=astextplain
|
*.doc diff=astextplain
|
||||||
*.DOC diff=astextplain
|
*.DOC diff=astextplain
|
||||||
*.docx diff=astextplain
|
*.docx diff=astextplain
|
||||||
*.DOCX diff=astextplain
|
*.DOCX diff=astextplain
|
||||||
*.dot diff=astextplain
|
*.dot diff=astextplain
|
||||||
*.DOT diff=astextplain
|
*.DOT diff=astextplain
|
||||||
*.pdf diff=astextplain
|
*.pdf diff=astextplain
|
||||||
*.PDF diff=astextplain
|
*.PDF diff=astextplain
|
||||||
*.rtf diff=astextplain
|
*.rtf diff=astextplain
|
||||||
*.RTF diff=astextplain
|
*.RTF diff=astextplain
|
||||||
*.md text
|
*.md text diff=markdown
|
||||||
*.tex text diff=tex
|
*.tex text diff=tex
|
||||||
*.adoc text
|
*.adoc text
|
||||||
*.textile text
|
*.textile text
|
||||||
@ -30,6 +30,7 @@
|
|||||||
*.tsv text
|
*.tsv text
|
||||||
*.txt text
|
*.txt text
|
||||||
*.sql text
|
*.sql text
|
||||||
|
*.ps1 text eol=crlf
|
||||||
|
|
||||||
# Graphics
|
# Graphics
|
||||||
*.png binary
|
*.png binary
|
||||||
@ -48,11 +49,28 @@
|
|||||||
|
|
||||||
# Scripts
|
# Scripts
|
||||||
*.bash text eol=lf
|
*.bash text eol=lf
|
||||||
|
*.fish text eol=lf
|
||||||
*.sh text eol=lf
|
*.sh text eol=lf
|
||||||
# These are explicitly windows files and should use crlf
|
# These are explicitly windows files and should use crlf
|
||||||
*.bat text eol=crlf
|
*.bat text eol=crlf
|
||||||
*.cmd text eol=crlf
|
*.cmd text eol=crlf
|
||||||
*.ps1 text eol=crlf
|
|
||||||
|
# web frontend stack -- force LF so SRI hashes are always correct
|
||||||
|
*.html text eol=lf
|
||||||
|
*.htm text eol=lf
|
||||||
|
*.css text eol=lf
|
||||||
|
*.min.css text eol=lf
|
||||||
|
*.js text eol=lf
|
||||||
|
*.min.js text eol=lf
|
||||||
|
*.php text eol=lf
|
||||||
|
|
||||||
|
# Visual Studio projects (Rider also)
|
||||||
|
*.cs diff=csharp
|
||||||
|
*.sln merge=union
|
||||||
|
*.csproj merge=union
|
||||||
|
*.vbproj merge=union
|
||||||
|
*.fsproj merge=union
|
||||||
|
*.dbproj merge=union
|
||||||
|
|
||||||
# Serialisation
|
# Serialisation
|
||||||
*.json text
|
*.json text
|
||||||
@ -65,11 +83,18 @@
|
|||||||
*.7z binary
|
*.7z binary
|
||||||
*.gz binary
|
*.gz binary
|
||||||
*.tar binary
|
*.tar binary
|
||||||
|
*.tgz binary
|
||||||
*.zip binary
|
*.zip binary
|
||||||
|
|
||||||
|
# Text files where line endings should be preserved
|
||||||
|
*.patch -text
|
||||||
|
|
||||||
#
|
#
|
||||||
# Exclude files from exporting
|
# Exclude files from exporting
|
||||||
#
|
#
|
||||||
|
|
||||||
.gitattributes export-ignore
|
.gitattributes export-ignore
|
||||||
.gitignore export-ignore
|
.gitignore export-ignore
|
||||||
|
.gitkeep export-ignore
|
||||||
|
.idea export-ignore
|
||||||
|
.vscode export-ignore
|
||||||
|
35
.gitignore
vendored
35
.gitignore
vendored
@ -1,10 +1,27 @@
|
|||||||
# ignore all vscode config files
|
#
|
||||||
.vscode/*
|
# JetBrains exclusions
|
||||||
!.vscode/settings.json
|
#
|
||||||
!.vscode/tasks.json
|
|
||||||
!.vscode/launch.json
|
|
||||||
!.vscode/extensions.json
|
|
||||||
!.vscode/numbered-bookmarks.json
|
|
||||||
|
|
||||||
# ignore all generated logfiles
|
# User-specific stuff
|
||||||
*.log
|
.idea/**/workspace.xml
|
||||||
|
.idea/**/tasks.xml
|
||||||
|
.idea/**/usage.statistics.xml
|
||||||
|
.idea/**/shelf
|
||||||
|
|
||||||
|
# Generated files
|
||||||
|
.idea/**/contentModel.xml
|
||||||
|
|
||||||
|
# Sensitive or high-churn files
|
||||||
|
.idea/**/dataSources/
|
||||||
|
.idea/**/dataSources.ids
|
||||||
|
.idea/**/dataSources.local.xml
|
||||||
|
.idea/**/sqlDataSources.xml
|
||||||
|
.idea/**/dynamic.xml
|
||||||
|
.idea/**/uiDesigner.xml
|
||||||
|
.idea/**/dbnavigator.xml
|
||||||
|
|
||||||
|
# modules
|
||||||
|
.idea_modules/
|
||||||
|
|
||||||
|
# Editor-based Rest Client
|
||||||
|
.idea/httpRequests
|
||||||
|
13
.idea/.idea.MailcowBackup.dir/.idea/.gitignore
vendored
Normal file
13
.idea/.idea.MailcowBackup.dir/.idea/.gitignore
vendored
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
# Default ignored files
|
||||||
|
/shelf/
|
||||||
|
/workspace.xml
|
||||||
|
# Rider ignored files
|
||||||
|
/.idea.MailcowBackup.iml
|
||||||
|
/projectSettingsUpdater.xml
|
||||||
|
/modules.xml
|
||||||
|
/contentModel.xml
|
||||||
|
# Datasource local storage ignored files
|
||||||
|
/../../../../../../../../../:\Users\AsifB\Documents\RiderProjects\MailcowBackup\.idea\.idea.MailcowBackup.dir\.idea/dataSources/
|
||||||
|
/dataSources.local.xml
|
||||||
|
# Editor-based HTTP Client requests
|
||||||
|
/httpRequests/
|
@ -0,0 +1,5 @@
|
|||||||
|
<component name="ProjectCodeStyleConfiguration">
|
||||||
|
<state>
|
||||||
|
<option name="PREFERRED_PROJECT_CODE_STYLE" value="Default" />
|
||||||
|
</state>
|
||||||
|
</component>
|
15
.idea/.idea.MailcowBackup.dir/.idea/git_toolbox_prj.xml
Normal file
15
.idea/.idea.MailcowBackup.dir/.idea/git_toolbox_prj.xml
Normal file
@ -0,0 +1,15 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<project version="4">
|
||||||
|
<component name="GitToolBoxProjectSettings">
|
||||||
|
<option name="commitMessageIssueKeyValidationOverride">
|
||||||
|
<BoolValueOverride>
|
||||||
|
<option name="enabled" value="true" />
|
||||||
|
</BoolValueOverride>
|
||||||
|
</option>
|
||||||
|
<option name="commitMessageValidationEnabledOverride">
|
||||||
|
<BoolValueOverride>
|
||||||
|
<option name="enabled" value="true" />
|
||||||
|
</BoolValueOverride>
|
||||||
|
</option>
|
||||||
|
</component>
|
||||||
|
</project>
|
8
.idea/.idea.MailcowBackup.dir/.idea/indexLayout.xml
Normal file
8
.idea/.idea.MailcowBackup.dir/.idea/indexLayout.xml
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<project version="4">
|
||||||
|
<component name="UserContentModel">
|
||||||
|
<attachedFolders />
|
||||||
|
<explicitIncludes />
|
||||||
|
<explicitExcludes />
|
||||||
|
</component>
|
||||||
|
</project>
|
23
.idea/.idea.MailcowBackup.dir/.idea/vcs.xml
Normal file
23
.idea/.idea.MailcowBackup.dir/.idea/vcs.xml
Normal file
@ -0,0 +1,23 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<project version="4">
|
||||||
|
<component name="CommitMessageInspectionProfile">
|
||||||
|
<profile version="1.0">
|
||||||
|
<inspection_tool class="BodyLimit" enabled="true" level="ERROR" enabled_by_default="true" />
|
||||||
|
<inspection_tool class="SubjectBodySeparation" enabled="true" level="ERROR" enabled_by_default="true" />
|
||||||
|
<inspection_tool class="SubjectLimit" enabled="true" level="WARNING" enabled_by_default="true">
|
||||||
|
<option name="RIGHT_MARGIN" value="50" />
|
||||||
|
</inspection_tool>
|
||||||
|
</profile>
|
||||||
|
</component>
|
||||||
|
<component name="GitSharedSettings">
|
||||||
|
<option name="FORCE_PUSH_PROHIBITED_PATTERNS">
|
||||||
|
<list>
|
||||||
|
<option value="master" />
|
||||||
|
<option value="main" />
|
||||||
|
</list>
|
||||||
|
</option>
|
||||||
|
</component>
|
||||||
|
<component name="VcsDirectoryMappings">
|
||||||
|
<mapping directory="$PROJECT_DIR$" vcs="Git" />
|
||||||
|
</component>
|
||||||
|
</project>
|
3
.vscode/numbered-bookmarks.json
vendored
3
.vscode/numbered-bookmarks.json
vendored
@ -1,3 +0,0 @@
|
|||||||
{
|
|
||||||
"bookmarks": []
|
|
||||||
}
|
|
147
LICENSE
147
LICENSE
@ -1,126 +1,21 @@
|
|||||||
GNU GENERAL PUBLIC LICENSE
|
MIT License
|
||||||
Version 3, 29 June 2007
|
|
||||||
Copyright © 2007 Free Software Foundation, Inc. <http s ://fsf.org/>
|
Copyright (c) 2020-2022 Asif Bacchus (asif@asifbacchus.dev)
|
||||||
Everyone is permitted to copy and distribute verbatim copies of this license document, but changing it is not allowed.
|
|
||||||
Preamble
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
The GNU General Public License is a free, copyleft license for software and other kinds of works.
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
The licenses for most software and other practical works are designed to take away your freedom to share and change the works. By contrast, the GNU General Public License is intended to guarantee your freedom to share and change all versions of a program--to make sure it remains free software for all its users. We, the Free Software Foundation, use the GNU General Public License for most of our software; it applies also to any other work released this way by its authors. You can apply it to your programs, too.
|
in the Software without restriction, including without limitation the rights
|
||||||
When we speak of free software, we are referring to freedom, not price. Our General Public Licenses are designed to make sure that you have the freedom to distribute copies of free software (and charge for them if you wish), that you receive source code or can get it if you want it, that you can change the software or use pieces of it in new free programs, and that you know you can do these things.
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
To protect your rights, we need to prevent others from denying you these rights or asking you to surrender the rights. Therefore, you have certain responsibilities if you distribute copies of the software, or if you modify it: responsibilities to respect the freedom of others.
|
copies of the Software, and to permit persons to whom the Software is
|
||||||
For example, if you distribute copies of such a program, whether gratis or for a fee, you must pass on to the recipients the same freedoms that you received. You must make sure that they, too, receive or can get the source code. And you must show them these terms so they know their rights.
|
furnished to do so, subject to the following conditions:
|
||||||
Developers that use the GNU GPL protect your rights with two steps: (1) assert copyright on the software, and (2) offer you this License giving you legal permission to copy, distribute and/or modify it.
|
|
||||||
For the developers' and authors' protection, the GPL clearly explains that there is no warranty for this free software. For both users' and authors' sake, the GPL requires that modified versions be marked as changed, so that their problems will not be attributed erroneously to authors of previous versions.
|
The above copyright notice and this permission notice shall be included in all
|
||||||
Some devices are designed to deny users access to install or run modified versions of the software inside them, although the manufacturer can do so. This is fundamentally incompatible with the aim of protecting users' freedom to change the software. The systematic pattern of such abuse occurs in the area of products for individuals to use, which is precisely where it is most unacceptable. Therefore, we have designed this version of the GPL to prohibit the practice for those products. If such problems arise substantially in other domains, we stand ready to extend this provision to those domains in future versions of the GPL, as needed to protect the freedom of users.
|
copies or substantial portions of the Software.
|
||||||
Finally, every program is threatened constantly by software patents. States should not allow patents to restrict development and use of software on general-purpose computers, but in those that do, we wish to avoid the special danger that patents applied to a free program could make it effectively proprietary. To prevent this, the GPL assures that patents cannot be used to render the program non-free.
|
|
||||||
The precise terms and conditions for copying, distribution and modification follow.
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
TERMS AND CONDITIONS
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
0. Definitions.
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
"This License" refers to version 3 of the GNU General Public License.
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
"Copyright" also means copyright-like laws that apply to other kinds of works, such as semiconductor masks.
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
"The Program" refers to any copyrightable work licensed under this License. Each licensee is addressed as "you". "Licensees" and "recipients" may be individuals or organizations.
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
To "modify" a work means to copy from or adapt all or part of the work in a fashion requiring copyright permission, other than the making of an exact copy. The resulting work is called a "modified version" of the earlier work or a work "based on" the earlier work.
|
SOFTWARE.
|
||||||
A "covered work" means either the unmodified Program or a work based on the Program.
|
|
||||||
To "propagate" a work means to do anything with it that, without permission, would make you directly or secondarily liable for infringement under applicable copyright law, except executing it on a computer or modifying a private copy. Propagation includes copying, distribution (with or without modification), making available to the public, and in some countries other activities as well.
|
|
||||||
To "convey" a work means any kind of propagation that enables other parties to make or receive copies. Mere interaction with a user through a computer network, with no transfer of a copy, is not conveying.
|
|
||||||
An interactive user interface displays "Appropriate Legal Notices" to the extent that it includes a convenient and prominently visible feature that (1) displays an appropriate copyright notice, and (2) tells the user that there is no warranty for the work (except to the extent that warranties are provided), that licensees may convey the work under this License, and how to view a copy of this License. If the interface presents a list of user commands or options, such as a menu, a prominent item in the list meets this criterion.
|
|
||||||
1. Source Code.
|
|
||||||
The "source code" for a work means the preferred form of the work for making modifications to it. "Object code" means any non-source form of a work.
|
|
||||||
A "Standard Interface" means an interface that either is an official standard defined by a recognized standards body, or, in the case of interfaces specified for a particular programming language, one that is widely used among developers working in that language.
|
|
||||||
The "System Libraries" of an executable work include anything, other than the work as a whole, that (a) is included in the normal form of packaging a Major Component, but which is not part of that Major Component, and (b) serves only to enable use of the work with that Major Component, or to implement a Standard Interface for which an implementation is available to the public in source code form. A "Major Component", in this context, means a major essential component (kernel, window system, and so on) of the specific operating system (if any) on which the executable work runs, or a compiler used to produce the work, or an object code interpreter used to run it.
|
|
||||||
The "Corresponding Source" for a work in object code form means all the source code needed to generate, install, and (for an executable work) run the object code and to modify the work, including scripts to control those activities. However, it does not include the work's System Libraries, or general-purpose tools or generally available free programs which are used unmodified in performing those activities but which are not part of the work. For example, Corresponding Source includes interface definition files associated with source files for the work, and the source code for shared libraries and dynamically linked subprograms that the work is specifically designed to require, such as by intimate data communication or control flow between those subprograms and other parts of the work.
|
|
||||||
The Corresponding Source need not include anything that users can regenerate automatically from other parts of the Corresponding Source.
|
|
||||||
The Corresponding Source for a work in source code form is that same work.
|
|
||||||
2. Basic Permissions.
|
|
||||||
All rights granted under this License are granted for the term of copyright on the Program, and are irrevocable provided the stated conditions are met. This License explicitly affirms your unlimited permission to run the unmodified Program. The output from running a covered work is covered by this License only if the output, given its content, constitutes a covered work. This License acknowledges your rights of fair use or other equivalent, as provided by copyright law.
|
|
||||||
You may make, run and propagate covered works that you do not convey, without conditions so long as your license otherwise remains in force. You may convey covered works to others for the sole purpose of having them make modifications exclusively for you, or provide you with facilities for running those works, provided that you comply with the terms of this License in conveying all material for which you do not control copyright. Those thus making or running the covered works for you must do so exclusively on your behalf, under your direction and control, on terms that prohibit them from making any copies of your copyrighted material outside their relationship with you.
|
|
||||||
Conveying under any other circumstances is permitted solely under the conditions stated below. Sublicensing is not allowed; section 10 makes it unnecessary.
|
|
||||||
3. Protecting Users' Legal Rights From Anti-Circumvention Law.
|
|
||||||
No covered work shall be deemed part of an effective technological measure under any applicable law fulfilling obligations under article 11 of the WIPO copyright treaty adopted on 20 December 1996, or similar laws prohibiting or restricting circumvention of such measures.
|
|
||||||
When you convey a covered work, you waive any legal power to forbid circumvention of technological measures to the extent such circumvention is effected by exercising rights under this License with respect to the covered work, and you disclaim any intention to limit operation or modification of the work as a means of enforcing, against the work's users, your or third parties' legal rights to forbid circumvention of technological measures.
|
|
||||||
4. Conveying Verbatim Copies.
|
|
||||||
You may convey verbatim copies of the Program's source code as you receive it, in any medium, provided that you conspicuously and appropriately publish on each copy an appropriate copyright notice; keep intact all notices stating that this License and any non-permissive terms added in accord with section 7 apply to the code; keep intact all notices of the absence of any warranty; and give all recipients a copy of this License along with the Program.
|
|
||||||
You may charge any price or no price for each copy that you convey, and you may offer support or warranty protection for a fee.
|
|
||||||
5. Conveying Modified Source Versions.
|
|
||||||
You may convey a work based on the Program, or the modifications to produce it from the Program, in the form of source code under the terms of section 4, provided that you also meet all of these conditions:
|
|
||||||
a) The work must carry prominent notices stating that you modified it, and giving a relevant date.
|
|
||||||
b) The work must carry prominent notices stating that it is released under this License and any conditions added under section 7. This requirement modifies the requirement in section 4 to "keep intact all notices".
|
|
||||||
c) You must license the entire work, as a whole, under this License to anyone who comes into possession of a copy. This License will therefore apply, along with any applicable section 7 additional terms, to the whole of the work, and all its parts, regardless of how they are packaged. This License gives no permission to license the work in any other way, but it does not invalidate such permission if you have separately received it.
|
|
||||||
d) If the work has interactive user interfaces, each must display Appropriate Legal Notices; however, if the Program has interactive interfaces that do not display Appropriate Legal Notices, your work need not make them do so.
|
|
||||||
A compilation of a covered work with other separate and independent works, which are not by their nature extensions of the covered work, and which are not combined with it such as to form a larger program, in or on a volume of a storage or distribution medium, is called an "aggregate" if the compilation and its resulting copyright are not used to limit the access or legal rights of the compilation's users beyond what the individual works permit. Inclusion of a covered work in an aggregate does not cause this License to apply to the other parts of the aggregate.
|
|
||||||
6. Conveying Non-Source Forms.
|
|
||||||
You may convey a covered work in object code form under the terms of sections 4 and 5, provided that you also convey the machine-readable Corresponding Source under the terms of this License, in one of these ways:
|
|
||||||
a) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by the Corresponding Source fixed on a durable physical medium customarily used for software interchange.
|
|
||||||
b) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by a written offer, valid for at least three years and valid for as long as you offer spare parts or customer support for that product model, to give anyone who possesses the object code either (1) a copy of the Corresponding Source for all the software in the product that is covered by this License, on a durable physical medium customarily used for software interchange, for a price no more than your reasonable cost of physically performing this conveying of source, or (2) access to copy the Corresponding Source from a network server at no charge.
|
|
||||||
c) Convey individual copies of the object code with a copy of the written offer to provide the Corresponding Source. This alternative is allowed only occasionally and noncommercially, and only if you received the object code with such an offer, in accord with subsection 6b.
|
|
||||||
d) Convey the object code by offering access from a designated place (gratis or for a charge), and offer equivalent access to the Corresponding Source in the same way through the same place at no further charge. You need not require recipients to copy the Corresponding Source along with the object code. If the place to copy the object code is a network server, the Corresponding Source may be on a different server (operated by you or a third party) that supports equivalent copying facilities, provided you maintain clear directions next to the object code saying where to find the Corresponding Source. Regardless of what server hosts the Corresponding Source, you remain obligated to ensure that it is available for as long as needed to satisfy these requirements.
|
|
||||||
e) Convey the object code using peer-to-peer transmission, provided you inform other peers where the object code and Corresponding Source of the work are being offered to the general public at no charge under subsection 6d.
|
|
||||||
A separable portion of the object code, whose source code is excluded from the Corresponding Source as a System Library, need not be included in conveying the object code work.
|
|
||||||
A "User Product" is either (1) a "consumer product", which means any tangible personal property which is normally used for personal, family, or household purposes, or (2) anything designed or sold for incorporation into a dwelling. In determining whether a product is a consumer product, doubtful cases shall be resolved in favor of coverage. For a particular product received by a particular user, "normally used" refers to a typical or common use of that class of product, regardless of the status of the particular user or of the way in which the particular user actually uses, or expects or is expected to use, the product. A product is a consumer product regardless of whether the product has substantial commercial, industrial or non-consumer uses, unless such uses represent the only significant mode of use of the product.
|
|
||||||
"Installation Information" for a User Product means any methods, procedures, authorization keys, or other information required to install and execute modified versions of a covered work in that User Product from a modified version of its Corresponding Source. The information must suffice to ensure that the continued functioning of the modified object code is in no case prevented or interfered with solely because modification has been made.
|
|
||||||
If you convey an object code work under this section in, or with, or specifically for use in, a User Product, and the conveying occurs as part of a transaction in which the right of possession and use of the User Product is transferred to the recipient in perpetuity or for a fixed term (regardless of how the transaction is characterized), the Corresponding Source conveyed under this section must be accompanied by the Installation Information. But this requirement does not apply if neither you nor any third party retains the ability to install modified object code on the User Product (for example, the work has been installed in ROM).
|
|
||||||
The requirement to provide Installation Information does not include a requirement to continue to provide support service, warranty, or updates for a work that has been modified or installed by the recipient, or for the User Product in which it has been modified or installed. Access to a network may be denied when the modification itself materially and adversely affects the operation of the network or violates the rules and protocols for communication across the network.
|
|
||||||
Corresponding Source conveyed, and Installation Information provided, in accord with this section must be in a format that is publicly documented (and with an implementation available to the public in source code form), and must require no special password or key for unpacking, reading or copying.
|
|
||||||
7. Additional Terms.
|
|
||||||
"Additional permissions" are terms that supplement the terms of this License by making exceptions from one or more of its conditions. Additional permissions that are applicable to the entire Program shall be treated as though they were included in this License, to the extent that they are valid under applicable law. If additional permissions apply only to part of the Program, that part may be used separately under those permissions, but the entire Program remains governed by this License without regard to the additional permissions.
|
|
||||||
When you convey a copy of a covered work, you may at your option remove any additional permissions from that copy, or from any part of it. (Additional permissions may be written to require their own removal in certain cases when you modify the work.) You may place additional permissions on material, added by you to a covered work, for which you have or can give appropriate copyright permission.
|
|
||||||
Notwithstanding any other provision of this License, for material you add to a covered work, you may (if authorized by the copyright holders of that material) supplement the terms of this License with terms:
|
|
||||||
a) Disclaiming warranty or limiting liability differently from the terms of sections 15 and 16 of this License; or
|
|
||||||
b) Requiring preservation of specified reasonable legal notices or author attributions in that material or in the Appropriate Legal Notices displayed by works containing it; or
|
|
||||||
c) Prohibiting misrepresentation of the origin of that material, or requiring that modified versions of such material be marked in reasonable ways as different from the original version; or
|
|
||||||
d) Limiting the use for publicity purposes of names of licensors or authors of the material; or
|
|
||||||
e) Declining to grant rights under trademark law for use of some trade names, trademarks, or service marks; or
|
|
||||||
f) Requiring indemnification of licensors and authors of that material by anyone who conveys the material (or modified versions of it) with contractual assumptions of liability to the recipient, for any liability that these contractual assumptions directly impose on those licensors and authors.
|
|
||||||
All other non-permissive additional terms are considered "further restrictions" within the meaning of section 10. If the Program as you received it, or any part of it, contains a notice stating that it is governed by this License along with a term that is a further restriction, you may remove that term. If a license document contains a further restriction but permits relicensing or conveying under this License, you may add to a covered work material governed by the terms of that license document, provided that the further restriction does not survive such relicensing or conveying.
|
|
||||||
If you add terms to a covered work in accord with this section, you must place, in the relevant source files, a statement of the additional terms that apply to those files, or a notice indicating where to find the applicable terms.
|
|
||||||
Additional terms, permissive or non-permissive, may be stated in the form of a separately written license, or stated as exceptions; the above requirements apply either way.
|
|
||||||
8. Termination.
|
|
||||||
You may not propagate or modify a covered work except as expressly provided under this License. Any attempt otherwise to propagate or modify it is void, and will automatically terminate your rights under this License (including any patent licenses granted under the third paragraph of section 11).
|
|
||||||
However, if you cease all violation of this License, then your license from a particular copyright holder is reinstated (a) provisionally, unless and until the copyright holder explicitly and finally terminates your license, and (b) permanently, if the copyright holder fails to notify you of the violation by some reasonable means prior to 60 days after the cessation.
|
|
||||||
Moreover, your license from a particular copyright holder is reinstated permanently if the copyright holder notifies you of the violation by some reasonable means, this is the first time you have received notice of violation of this License (for any work) from that copyright holder, and you cure the violation prior to 30 days after your receipt of the notice.
|
|
||||||
Termination of your rights under this section does not terminate the licenses of parties who have received copies or rights from you under this License. If your rights have been terminated and not permanently reinstated, you do not qualify to receive new licenses for the same material under section 10.
|
|
||||||
9. Acceptance Not Required for Having Copies.
|
|
||||||
You are not required to accept this License in order to receive or run a copy of the Program. Ancillary propagation of a covered work occurring solely as a consequence of using peer-to-peer transmission to receive a copy likewise does not require acceptance. However, nothing other than this License grants you permission to propagate or modify any covered work. These actions infringe copyright if you do not accept this License. Therefore, by modifying or propagating a covered work, you indicate your acceptance of this License to do so.
|
|
||||||
10. Automatic Licensing of Downstream Recipients.
|
|
||||||
Each time you convey a covered work, the recipient automatically receives a license from the original licensors, to run, modify and propagate that work, subject to this License. You are not responsible for enforcing compliance by third parties with this License.
|
|
||||||
An "entity transaction" is a transaction transferring control of an organization, or substantially all assets of one, or subdividing an organization, or merging organizations. If propagation of a covered work results from an entity transaction, each party to that transaction who receives a copy of the work also receives whatever licenses to the work the party's predecessor in interest had or could give under the previous paragraph, plus a right to possession of the Corresponding Source of the work from the predecessor in interest, if the predecessor has it or can get it with reasonable efforts.
|
|
||||||
You may not impose any further restrictions on the exercise of the rights granted or affirmed under this License. For example, you may not impose a license fee, royalty, or other charge for exercise of rights granted under this License, and you may not initiate litigation (including a cross-claim or counterclaim in a lawsuit) alleging that any patent claim is infringed by making, using, selling, offering for sale, or importing the Program or any portion of it.
|
|
||||||
11. Patents.
|
|
||||||
A "contributor" is a copyright holder who authorizes use under this License of the Program or a work on which the Program is based. The work thus licensed is called the contributor's "contributor version".
|
|
||||||
A contributor's "essential patent claims" are all patent claims owned or controlled by the contributor, whether already acquired or hereafter acquired, that would be infringed by some manner, permitted by this License, of making, using, or selling its contributor version, but do not include claims that would be infringed only as a consequence of further modification of the contributor version. For purposes of this definition, "control" includes the right to grant patent sublicenses in a manner consistent with the requirements of this License.
|
|
||||||
Each contributor grants you a non-exclusive, worldwide, royalty-free patent license under the contributor's essential patent claims, to make, use, sell, offer for sale, import and otherwise run, modify and propagate the contents of its contributor version.
|
|
||||||
In the following three paragraphs, a "patent license" is any express agreement or commitment, however denominated, not to enforce a patent (such as an express permission to practice a patent or covenant not to sue for patent infringement). To "grant" such a patent license to a party means to make such an agreement or commitment not to enforce a patent against the party.
|
|
||||||
If you convey a covered work, knowingly relying on a patent license, and the Corresponding Source of the work is not available for anyone to copy, free of charge and under the terms of this License, through a publicly available network server or other readily accessible means, then you must either (1) cause the Corresponding Source to be so available, or (2) arrange to deprive yourself of the benefit of the patent license for this particular work, or (3) arrange, in a manner consistent with the requirements of this License, to extend the patent license to downstream recipients. "Knowingly relying" means you have actual knowledge that, but for the patent license, your conveying the covered work in a country, or your recipient's use of the covered work in a country, would infringe one or more identifiable patents in that country that you have reason to believe are valid.
|
|
||||||
If, pursuant to or in connection with a single transaction or arrangement, you convey, or propagate by procuring conveyance of, a covered work, and grant a patent license to some of the parties receiving the covered work authorizing them to use, propagate, modify or convey a specific copy of the covered work, then the patent license you grant is automatically extended to all recipients of the covered work and works based on it.
|
|
||||||
A patent license is "discriminatory" if it does not include within the scope of its coverage, prohibits the exercise of, or is conditioned on the non-exercise of one or more of the rights that are specifically granted under this License. You may not convey a covered work if you are a party to an arrangement with a third party that is in the business of distributing software, under which you make payment to the third party based on the extent of your activity of conveying the work, and under which the third party grants, to any of the parties who would receive the covered work from you, a discriminatory patent license (a) in connection with copies of the covered work conveyed by you (or copies made from those copies), or (b) primarily for and in connection with specific products or compilations that contain the covered work, unless you entered into that arrangement, or that patent license was granted, prior to 28 March 2007.
|
|
||||||
Nothing in this License shall be construed as excluding or limiting any implied license or other defenses to infringement that may otherwise be available to you under applicable patent law.
|
|
||||||
12. No Surrender of Others' Freedom.
|
|
||||||
If conditions are imposed on you (whether by court order, agreement or otherwise) that contradict the conditions of this License, they do not excuse you from the conditions of this License. If you cannot convey a covered work so as to satisfy simultaneously your obligations under this License and any other pertinent obligations, then as a consequence you may not convey it at all. For example, if you agree to terms that obligate you to collect a royalty for further conveying from those to whom you convey the Program, the only way you could satisfy both those terms and this License would be to refrain entirely from conveying the Program.
|
|
||||||
13. Use with the GNU Affero General Public License.
|
|
||||||
Notwithstanding any other provision of this License, you have permission to link or combine any covered work with a work licensed under version 3 of the GNU Affero General Public License into a single combined work, and to convey the resulting work. The terms of this License will continue to apply to the part which is the covered work, but the special requirements of the GNU Affero General Public License, section 13, concerning interaction through a network will apply to the combination as such.
|
|
||||||
14. Revised Versions of this License.
|
|
||||||
The Free Software Foundation may publish revised and/or new versions of the GNU General Public License from time to time. Such new versions will be similar in spirit to the present version, but may differ in detail to address new problems or concerns.
|
|
||||||
Each version is given a distinguishing version number. If the Program specifies that a certain numbered version of the GNU General Public License "or any later version" applies to it, you have the option of following the terms and conditions either of that numbered version or of any later version published by the Free Software Foundation. If the Program does not specify a version number of the GNU General Public License, you may choose any version ever published by the Free Software Foundation.
|
|
||||||
If the Program specifies that a proxy can decide which future versions of the GNU General Public License can be used, that proxy's public statement of acceptance of a version permanently authorizes you to choose that version for the Program.
|
|
||||||
Later license versions may give you additional or different permissions. However, no additional obligations are imposed on any author or copyright holder as a result of your choosing to follow a later version.
|
|
||||||
15. Disclaimer of Warranty.
|
|
||||||
THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION.
|
|
||||||
16. Limitation of Liability.
|
|
||||||
IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
|
|
||||||
17. Interpretation of Sections 15 and 16.
|
|
||||||
If the disclaimer of warranty and limitation of liability provided above cannot be given local legal effect according to their terms, reviewing courts shall apply local law that most closely approximates an absolute waiver of all civil liability in connection with the Program, unless a warranty or assumption of liability accompanies a copy of the Program in return for a fee. END OF TERMS AND CONDITIONS
|
|
||||||
How to Apply These Terms to Your New Programs
|
|
||||||
If you develop a new program, and you want it to be of the greatest possible use to the public, the best way to achieve this is to make it free software which everyone can redistribute and change under these terms.
|
|
||||||
To do so, attach the following notices to the program. It is safest to attach them to the start of each source file to most effectively state the exclusion of warranty; and each file should have at least the "copyright" line and a pointer to where the full notice is found.
|
|
||||||
<one line to give the program's name and a brief idea of what it does.>
|
|
||||||
Copyright (C) <year> <name of author>
|
|
||||||
This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.
|
|
||||||
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
||||||
You should have received a copy of the GNU General Public License along with this program. If not, see <http s ://www.gnu.org/licenses/>.
|
|
||||||
Also add information on how to contact you by electronic and paper mail.
|
|
||||||
If the program does terminal interaction, make it output a short notice like this when it starts in an interactive mode:
|
|
||||||
<program> Copyright (C) <year> <name of author>
|
|
||||||
This program comes with ABSOLUTELY NO WARRANTY; for details type `show w'.
|
|
||||||
This is free software, and you are welcome to redistribute it under certain conditions; type `show c' for details.
|
|
||||||
The hypothetical commands `show w' and `show c' should show the appropriate parts of the General Public License. Of course, your program's commands might be different; for a GUI interface, you would use an "about box".
|
|
||||||
You should also get your employer (if you work as a programmer) or school, if any, to sign a "copyright disclaimer" for the program, if necessary. For more information on this, and how to apply and follow the GNU GPL, see <http s ://www.gnu.org/licenses/>.
|
|
||||||
The GNU General Public License does not permit incorporating your program into proprietary programs. If your program is a subroutine library, you may consider it more useful to permit linking proprietary applications with the library. If this is what you want to do, use the GNU Lesser General Public License instead of this License. But first, please read <http s ://www.gnu.org/philosophy/why-not-lgpl.html>.
|
|
||||||
|
596
README.md
596
README.md
@ -1,569 +1,129 @@
|
|||||||
# Mailcow Backup Using borgbackup <!-- omit in toc -->
|
# Mailcow Backup Using borgbackup
|
||||||
|
|
||||||
This script automates backing up your Mailcow installation using borgbackup
|
This script automates backing up your Mailcow installation using borgbackup and a remote ssh-capable storage system. I suggest using rsync.net since they
|
||||||
and a remote ssh-capable storage system. I suggest using rsync.net since they
|
have great speeds and a special pricing structure for borgbackup/attic users ([details here](https://www.rsync.net/products/attic.html)).
|
||||||
have great speeds and a special pricing structure for borgbackup/attic users
|
|
||||||
([details here](https://www.rsync.net/products/attic.html)).
|
|
||||||
|
|
||||||
This script automates the following tasks:
|
This script automates the following tasks:
|
||||||
|
|
||||||
- Optionally copies a 503 error page to your webserver so users know when your
|
- Optionally copies a 503 error page to your webserver so users know when your server is unavailable due to backups being performed. The 503 file is removed
|
||||||
server is unavailable due to backups being performed. The 503 file is removed
|
|
||||||
when the backup is completed so users can login again
|
when the backup is completed so users can login again
|
||||||
- Dumps the Mailcow mySQL database and adds it to the backup
|
- Dumps the Mailcow mySQL database and adds it to the backup
|
||||||
- Handles stopping and re-starting mail-flow containers (postfix and dovecot) so
|
- Handles stopping and re-starting mail-flow containers (postfix and dovecot) so everything is in a consistent state during the backup
|
||||||
everything is in a consistent state during the backup
|
|
||||||
- Allows you to specify additional files you want backed up
|
- Allows you to specify additional files you want backed up
|
||||||
- Allows you to specify files/directories to exclude from your backups
|
- Allows you to specify files/directories to exclude from your backups
|
||||||
- Runs 'borg prune' to make sure you are trimming old backups on your schedule
|
- Runs 'borg prune' to make sure you are trimming old backups on your schedule
|
||||||
- Creates a clear, easy to parse log file so you can keep an eye on your backups
|
- Creates a clear, easy to parse log file so you can keep an eye on your backups and any errors/warnings
|
||||||
and any errors/warnings
|
|
||||||
|
|
||||||
## Contents <!-- omit in toc -->
|
## contents
|
||||||
|
|
||||||
- [Installation/copying](#installationcopying)
|
<!-- toc -->
|
||||||
- [Environment notes](#environment-notes)
|
|
||||||
- [Why this script must be run as root](#why-this-script-must-be-run-as-root)
|
|
||||||
- [Script parameters](#script-parameters)
|
|
||||||
- [Optional parameters](#optional-parameters)
|
|
||||||
- [Docker container STOP timeout before error: -1 _number_](#docker-container-stop-timeout-before-error--1-number)
|
|
||||||
- [Docker container START timeout before error: -2 _number_](#docker-container-start-timeout-before-error--2-number)
|
|
||||||
- [Path to 503 error page: -5 _/path/to/filename.html_](#path-to-503-error-page--5-pathtofilenamehtml)
|
|
||||||
- [Path to borg details file: -b _/path/to/filename.file_](#path-to-borg-details-file--b-pathtofilenamefile)
|
|
||||||
- [File name of docker-compose configuration file: -d _filename.file_](#file-name-of-docker-compose-configuration-file--d-filenamefile)
|
|
||||||
- [Log file location: -l _/path/to/filename.file_](#log-file-location--l-pathtofilenamefile)
|
|
||||||
- [File name of Mailcow master configuration file: -m _filename.file_](#file-name-of-mailcow-master-configuration-file--m-filenamefile)
|
|
||||||
- [Verbose output from borg: -v (no arguments)](#verbose-output-from-borg--v-no-arguments)
|
|
||||||
- [Path to webroot: -w _/path/to/webroot/_](#path-to-webroot--w-pathtowebroot)
|
|
||||||
- [Borg details file](#borg-details-file)
|
|
||||||
- [Protect your borg details file](#protect-your-borg-details-file)
|
|
||||||
- [borg specific entries (lines 1-4)](#borg-specific-entries-lines-1-4)
|
|
||||||
- [Line 1: Path to borg base directory](#line-1-path-to-borg-base-directory)
|
|
||||||
- [Line 2: Path to SSH key for remote server](#line-2-path-to-ssh-key-for-remote-server)
|
|
||||||
- [Line 3: Connection string to remote repo](#line-3-connection-string-to-remote-repo)
|
|
||||||
- [Line 4: Password for borg repo/repo key](#line-4-password-for-borg-reporepo-key)
|
|
||||||
- [additional files/directories to backup](#additional-filesdirectories-to-backup)
|
|
||||||
- [exclusion patterns](#exclusion-patterns)
|
|
||||||
- [prune timeframe options](#prune-timeframe-options)
|
|
||||||
- [borg remote location](#borg-remote-location)
|
|
||||||
- [Examples](#examples)
|
|
||||||
- [503 functionality](#503-functionality)
|
|
||||||
- [Conditional forwarding by your webserver](#conditional-forwarding-by-your-webserver)
|
|
||||||
- [NGINX](#nginx)
|
|
||||||
- [Apache](#apache)
|
|
||||||
- [Disabling 503 functionality altogether](#disabling-503-functionality-altogether)
|
|
||||||
- [Scheduling: Cron](#scheduling-cron)
|
|
||||||
- [The log file](#the-log-file)
|
|
||||||
- [Using Logwatch](#using-logwatch)
|
|
||||||
- [Remember to rotate your logs](#remember-to-rotate-your-logs)
|
|
||||||
- [Final notes](#final-notes)
|
|
||||||
|
|
||||||
## Installation/copying
|
- [quick start](#quick-start)
|
||||||
|
- [configuration file](#configuration-file)
|
||||||
|
- [running the script](#running-the-script)
|
||||||
|
- [scheduling your backup via cron](#scheduling-your-backup-via-cron)
|
||||||
|
- [restoring backups](#restoring-backups)
|
||||||
|
- [final notes](#final-notes)
|
||||||
|
|
||||||
Once you've either cloned this git or downloaded the release file, simply copy
|
<!-- tocstop -->
|
||||||
the files within the archive to whatever location(s) that work for your setup.
|
|
||||||
I've stored the files in this git archive in a directory structure that should
|
|
||||||
match most default setups. I suggest keeping the contents of the
|
|
||||||
*'/root/scripts'* folder in that location since the root user must execute the
|
|
||||||
script anyways. If you edit the 503.html and mc_borg.details files in place,
|
|
||||||
then you don't have to specify their locations when running the script.
|
|
||||||
|
|
||||||
Remember to make the script executable!
|
## quick start
|
||||||
|
|
||||||
```Bash
|
Clone this repo or download a release file into a directory of your choosing. For all examples in this document, I will assume you will run the script from */scripts/backup*. Make sure the script file is executable and you protect the *.details* file since it contains things like your repo password:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# run commands as root
|
||||||
|
sudo -s
|
||||||
|
|
||||||
|
# find somewhere to clone the repo
|
||||||
|
cd /usr/local/src
|
||||||
|
|
||||||
|
# clone the repo from my server (best choice)
|
||||||
|
git clone https://git.asifbacchus.app/asif/MailcowBackup.git
|
||||||
|
# or clone from github
|
||||||
|
git clone https://github.com/asifbacchus/MailcowBackup.git
|
||||||
|
|
||||||
|
# make a home for your backup script
|
||||||
|
mkdir -p /scripts/backup
|
||||||
|
cd /scripts/backup
|
||||||
|
|
||||||
|
# copy files from cloned repo to this new home
|
||||||
|
cp /usr/local/src/MailcowBackup/backup/* ./
|
||||||
|
|
||||||
|
# make script executable and protect your .details file
|
||||||
chmod +x backup.sh
|
chmod +x backup.sh
|
||||||
|
chmod 600 backup.details
|
||||||
```
|
```
|
||||||
|
|
||||||
In addition, you can rename this script file to anything you like. The log file
|
## configuration file
|
||||||
will use that same name by default when naming itself and any mention of this
|
|
||||||
file in the logs will automatically use whatever name you choose to give it.
|
|
||||||
|
|
||||||
## Environment notes
|
You will need to let the script know how to access your remote repo along with any passwords/keyfiles needed to encrypt data. This is all handled via the plain-text 'configuration details' file. By default, this file is named *backup.details*. The file itself is fully commented so setting it up should not be difficult. If you need more information, consult [page 4.0](https://git.asifbacchus.app/asif/MailcowBackup/wiki/4.0-Configuration-details-file) in the wiki.
|
||||||
|
|
||||||
The script is designed to be easy to use but still be flexible enough to
|
## running the script
|
||||||
accommodate a wide range of Mailcow setups. The script pulls nearly all it's
|
|
||||||
configuration from the Mailcow configuration files themselves, so it adapts to
|
|
||||||
nearly all customizations you may have in your environment. The script accepts
|
|
||||||
several optional parameters to override its default or detected settings. In
|
|
||||||
addition, it reads easy-to-edit external plain-text files for borg settings so
|
|
||||||
you don't have to weed through the script code to supply things like passwords.
|
|
||||||
|
|
||||||
**This script auto-detects the location of your Mailcow configuration file. If
|
After setting up the *.details* file correctly and assuming you are running a default set up of mailcow according to the documentation, you just have to run the script and it will find everything on it's own. In particular, the defaults are set as follows:
|
||||||
you have multiple files on your system with the same name as the configuration
|
|
||||||
file, the script will likely get confused and exit with an error**
|
|
||||||
|
|
||||||
## Why this script must be run as root
|
- mailcow.conf is located at */opt/mailcow-dockerized/mailcow.conf*
|
||||||
|
- docker-compose file is located at */opt/mailcow-dockerized/docker-compose.yml*
|
||||||
|
- the log file will be saved in the same directory as the script with the same name as the script but with the extension *.log*
|
||||||
|
|
||||||
This script must be run by the root user and will exit with an error if you try
|
To get a list of all configuration options with defaults:
|
||||||
running it otherwise. This is because a default secured setup of borgbackup
|
|
||||||
contains things like the repository private key that are locked out to root user
|
|
||||||
access only. In addition, the root user is guaranteed to have access to all
|
|
||||||
files you might want to backup.
|
|
||||||
|
|
||||||
## Script parameters
|
```bash
|
||||||
|
./backup.sh --help
|
||||||
You can run the script with the *'-?'* parameter to access the built-in help
|
|
||||||
which explains the parameters. However, the following is a more detailed
|
|
||||||
explanation of each parameter and how to use them. **Note that any parameters
|
|
||||||
needing a directory (webroot, log file location, etc.) can be entered with or
|
|
||||||
without the trailing '/' since it's stripped by the script anyways.**
|
|
||||||
|
|
||||||
General usage:
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
/path/to/script/scriptname.sh -parameter argument -parameter argument ...
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Optional parameters
|
To run with defaults:
|
||||||
|
|
||||||
#### Docker container STOP timeout before error: -1 _number_
|
```bash
|
||||||
|
./backup.sh
|
||||||
The amount of time, in seconds, to wait for a docker container to STOP
|
|
||||||
gracefully before aborting, logging the error and exiting the script.\
|
|
||||||
**Default: _120_**
|
|
||||||
|
|
||||||
#### Docker container START timeout before error: -2 _number_
|
|
||||||
|
|
||||||
The amount of time, in seconds, to wait for a docker container to START
|
|
||||||
before aborting, logging the error and exiting the script.\
|
|
||||||
**Default: _180_**
|
|
||||||
|
|
||||||
#### Path to 503 error page: -5 _/path/to/filename.html_
|
|
||||||
|
|
||||||
The path to an html file for the script to copy to your webroot during the
|
|
||||||
backup process. This file can be scanned by your webserver and a 503 error can
|
|
||||||
be issued to users letting them know that your Mailcow is 'temporarily
|
|
||||||
unavailable' while being backed up. A sample 503 page is included for you.
|
|
||||||
|
|
||||||
If you remove the default file or the one you specify is missing, a warning will
|
|
||||||
be issued by the script but, it will continue executing. More details on the
|
|
||||||
503 notification can be found later in the [503
|
|
||||||
functionality](#503-functionality) section of this document.\
|
|
||||||
**Default: _scriptpath/503.html_**
|
|
||||||
|
|
||||||
#### Path to borg details file: -b _/path/to/filename.file_
|
|
||||||
|
|
||||||
This is a text file that lays out various borg options such as repo name,
|
|
||||||
password, additional files to include, exclusion patters, etc. A sample file is
|
|
||||||
included for your reference. More details, including the *required order* of
|
|
||||||
entries can be found later in this document in the [borg details
|
|
||||||
file](#borg-details-file) section.\
|
|
||||||
**Default: _scriptpath/mc_borg.details_**
|
|
||||||
|
|
||||||
#### File name of docker-compose configuration file: -d _filename.file_
|
|
||||||
|
|
||||||
This is the file name of your docker-compose configuration file that is used to
|
|
||||||
build/start/stop containers. This script will only search for this file within
|
|
||||||
the same directory where your Mailcow configuration file is found.\
|
|
||||||
**Default: _docker-compose.yml_**
|
|
||||||
|
|
||||||
#### Log file location: -l _/path/to/filename.file_
|
|
||||||
|
|
||||||
If you have a particular place and filename you'd like this script to use for
|
|
||||||
it's log, then you can specify it using this parameter. I would recommend
|
|
||||||
*'/var/log/backup.log'*. By default, the script will name the log file
|
|
||||||
*scriptname*.log and will save it in the same directory as the script itself.\
|
|
||||||
**Default: _scriptpath/scriptname.log_**
|
|
||||||
|
|
||||||
#### File name of Mailcow master configuration file: -m _filename.file_
|
|
||||||
|
|
||||||
This is the file name of the Mailcow master configuration file that was
|
|
||||||
generated after installation and contains all information needed to run Mailcow
|
|
||||||
(database user name, volume directory prefixes, etc.) This script will search
|
|
||||||
your computer for either the default file name or the one you have provided.
|
|
||||||
Upon finding it, the script will derive the file path and use that as the path
|
|
||||||
in which to run all Mailcow/docker commands. **Please do not have multiple
|
|
||||||
files on your system with this name, the script WILL get confused and exit with
|
|
||||||
an error.\
|
|
||||||
**Default: _mailcow.conf_**
|
|
||||||
|
|
||||||
#### Verbose output from borg: -v (no arguments)
|
|
||||||
|
|
||||||
By default, the script will ask borg to generate summary only output and record
|
|
||||||
that in the script's log file. If you are running the backup for the first time
|
|
||||||
or are troubleshooting, you may want a detailed output of all files and their
|
|
||||||
changed/unchanged/excluded status from borg. In that case, specify the -v
|
|
||||||
switch. **Note: This will make your log file very large very quickly since EVERY
|
|
||||||
file being backed up is written to the log.**
|
|
||||||
|
|
||||||
#### Path to webroot: -w _/path/to/webroot/_
|
|
||||||
|
|
||||||
This is the path to the directory your webserver is using as it's default root.
|
|
||||||
In other words, this is the directory that contains the html files served when
|
|
||||||
someone browses to your server. The correct webroot depends greatly on your
|
|
||||||
particular setup.
|
|
||||||
|
|
||||||
If you directly connect to Mailcow via Docker, then your webroot is by default
|
|
||||||
*/opt/mailcow-dockerized/data/web*, unless you've made changes to your install
|
|
||||||
locations. If you are running behind a reverse-proxy, then your webroot is your
|
|
||||||
webserver's webroot (*/var/www* or */usr/share/nginx/html*, for example).
|
|
||||||
|
|
||||||
This is used exclusively for 503 functionality since the script has to know
|
|
||||||
where to copy the 503 file. If you don't want to use this functionality, you
|
|
||||||
can omit this parameter and the script will issue a warning and move on. More
|
|
||||||
details can be found in the [503 functionality](#503-functionality) section
|
|
||||||
later in this document.
|
|
||||||
|
|
||||||
## Borg details file
|
|
||||||
|
|
||||||
This file contains all the data needed to access your borg remote data repo.
|
|
||||||
Each line must contain specific information in a specific order or **needs to be
|
|
||||||
blank if that data is not required**. The sample file includes this data and
|
|
||||||
example entries. The file must have the following information in the following
|
|
||||||
order:
|
|
||||||
|
|
||||||
1. path to borg base directory **(required)**
|
|
||||||
2. path to ssh private key for remote server **(required)**
|
|
||||||
3. connection string to remote repo **(required)**
|
|
||||||
4. password for borg repo/repo key **(required)**
|
|
||||||
5. path to file listing additional files/directories to backup
|
|
||||||
6. path to file containing borg-specific exclusion patterns
|
|
||||||
7. prune timeframe options
|
|
||||||
8. location of borg remote instance
|
|
||||||
|
|
||||||
### Protect your borg details file
|
|
||||||
|
|
||||||
This file contains information on how to access and decrypt your borg repo,
|
|
||||||
therefore, you **must** protect it. You should lock it out for everyone but
|
|
||||||
your root user. Putting it in your root folder is not enough! Run the following
|
|
||||||
commands to restrict access to the root user only (assuming filename is
|
|
||||||
*mc_borg.details*):
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
chown root:root mc_borg.details # make root the owner of this file
|
|
||||||
chmod 600 mc_borg.details # grant access to root user only (read/write)
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### borg specific entries (lines 1-4)
|
To run with a custom log file name and location:
|
||||||
|
|
||||||
If you need help with these options, then you should consult the borg
|
```bash
|
||||||
documentation or search my blog at
|
./backup.sh --log /var/log/mailcow_backup.log
|
||||||
[https://mytechiethoughts.com](https://mytechiethoughts.com) for borg. Here's a
|
|
||||||
very brief overview:
|
|
||||||
|
|
||||||
#### Line 1: Path to borg base directory
|
|
||||||
|
|
||||||
This is primary directory on your local system where your borg configuration is
|
|
||||||
located, **NOT* the path to your borg binary. The base directory contains the
|
|
||||||
borg configuration, cache, security files and keys.
|
|
||||||
|
|
||||||
#### Line 2: Path to SSH key for remote server
|
|
||||||
|
|
||||||
This is the SSH key used to connect to your remote (backup) server where your
|
|
||||||
borg repo is located. **This is NOT your borg repo key!**
|
|
||||||
|
|
||||||
> Please note: If you are planning on executing this script via cron or some
|
|
||||||
> other form of automation, it is *highly recommended* that you use an SSH key
|
|
||||||
> **without** a password! SSH is designed such that passwords cannot simply be
|
|
||||||
> passed to it via environment variables, etc. so this is something not easily
|
|
||||||
> automated by a script such as this for security reasons. As such, your
|
|
||||||
> computer will sit and wait for you to enter the password and will NOT execute
|
|
||||||
> the actual backup portion of the script until the SSH key password is provided.
|
|
||||||
>
|
|
||||||
> If you really want/need to use an SSH key password, you will have to look into
|
|
||||||
> somethign like GNOME keyring or SSH-agent to provide a secure automated way to
|
|
||||||
> provide that password to SSH and allow this script to continue.
|
|
||||||
>
|
|
||||||
> In practice, SSH keys without passwords are still quite safe since the key
|
|
||||||
> must still be known in order to connect and most keys are quite long. In
|
|
||||||
> addition, they key only connects to the remote server, your actual information
|
|
||||||
> within the borg repository is still encrypted and secured with both a key and
|
|
||||||
> password.
|
|
||||||
|
|
||||||
#### Line 3: Connection string to remote repo
|
|
||||||
|
|
||||||
This is the full server and path required to connect to your borg repo on the
|
|
||||||
remote server. Very often it is the in the form of:
|
|
||||||
|
|
||||||
```
|
|
||||||
user@servername.tld:repo-name/
|
|
||||||
```
|
```
|
||||||
|
|
||||||
for rsync.net it is in the following form:
|
To copy a 503 error page to your webroot:
|
||||||
|
|
||||||
```
|
```bash
|
||||||
username@server-number.rsync.net:repo-name/
|
# assuming default NGINX webroot (/usr/share/nginx/html)
|
||||||
|
./backup.sh -5
|
||||||
|
# custom webroot
|
||||||
|
./backup.sh -5 -w /var/www/
|
||||||
```
|
```
|
||||||
|
|
||||||
#### Line 4: Password for borg repo/repo key
|
Common usage: custom log file and copy 503 to custom webroot
|
||||||
|
|
||||||
This is the password needed to access and decrypt your *borg repo*. Assuming
|
```bash
|
||||||
you set up your borg repo using recommended practices, this will actually be the
|
./backup.sh -l /var/log/mailcow_backup.log -5 -w /var/www/
|
||||||
password for your *borg repo private key*. **This is NOT your SSH key
|
|
||||||
password!**
|
|
||||||
|
|
||||||
### additional files/directories to backup
|
|
||||||
|
|
||||||
This points to a plain-text file listing additional files and directories you'd
|
|
||||||
like borg to include in the backup. The sample file, *'xtraLocations.borg'*
|
|
||||||
contains the most likely files you'd want to include assuming you're using a
|
|
||||||
standard setup like I outline in my blog.
|
|
||||||
|
|
||||||
The following would include all files in the home folder for users *'foo'* and
|
|
||||||
*'bar'* and any conf files in *'/etc/someProgram'*:
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
/home/foo/
|
|
||||||
/home/bar/
|
|
||||||
/etc/someProgram/*.conf
|
|
||||||
```
|
```
|
||||||
|
|
||||||
*You can leave this line blank* to tell borg to only backup your Mailcow data,
|
Non-default mailcow location (example: */var/mailcow*):
|
||||||
configuration and the SQL dump. However, this is pretty unusual since you would
|
|
||||||
not be including any server configuration files, reverse-proxy configurations,
|
|
||||||
etc. If you omit this line, the script will log a warning to remind you of this
|
|
||||||
unusual situation.
|
|
||||||
|
|
||||||
### exclusion patterns
|
```bash
|
||||||
|
./backup.sh --docker-compose /var/mailcow/docker-compose.yml --mailcow-config /var/mailcow/mailcow.conf
|
||||||
This points to a plain-text file containing borg-specific patterns describing
|
|
||||||
what files you'd like borg to ignore during the backup. To specify exclusions,
|
|
||||||
create a text file in any location you want and specify exclusions patterns, one
|
|
||||||
per line. Then update line 6 in your borg details file with the path to your
|
|
||||||
new exclusion file.
|
|
||||||
|
|
||||||
You need to run *'borg help patterns'* for help on how to specify exclusion
|
|
||||||
patterns since the format is not always standard BASH format and only sometimes
|
|
||||||
uses standard regex.
|
|
||||||
|
|
||||||
If you leave this line blank, the script will note it is not processing any
|
|
||||||
exclusions and will proceed with backing up all files specified.
|
|
||||||
|
|
||||||
### prune timeframe options
|
|
||||||
|
|
||||||
Here you can let borg prune know how you want to manage your backup history.
|
|
||||||
Consult the borg documentation and then copy the relevant options directly into
|
|
||||||
this line including any spaces, etc. The example file contains the following as
|
|
||||||
a staring point:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
--keep-within=7d --keep-daily=30 --keep-weekly=12 --keep-monthly=-1
|
|
||||||
```
|
```
|
||||||
|
|
||||||
This would tell borg prune to keep ALL backups made for any reason within the
|
For more configuration options, see [page 3.0](https://git.asifbacchus.app/asif/MailcowBackup/wiki/3.0-Script-parameters) in the wiki and [page 4.4](https://git.asifbacchus.app/asif/MailcowBackup/wiki/4.4-Configuration-examples) for some configuration examples. Consult [section 7](https://git.asifbacchus.app/asif/MailcowBackup/wiki/7.0-Logs) of the wiki for information about the log file and how to integrate it with logwatch.
|
||||||
last 7 days, keep 30 days worth of daily backups, 12 weeks of end-of-week
|
|
||||||
backups and then an infinite amount of end-of-month backups.
|
|
||||||
|
|
||||||
### borg remote location
|
## scheduling your backup via cron
|
||||||
|
|
||||||
If you're using rsync.net, then just have this say *'borg1'*. If you are using
|
Edit your root user's crontab and add an entry like this which would run the script using defaults at 1:07am daily:
|
||||||
another provider, you'll have to reference their locally installed copy of borg
|
|
||||||
relative to your home directory. You can also leave this blank if your provider
|
|
||||||
does not run borg locally but your backups/restores will be slower.
|
|
||||||
|
|
||||||
### Examples
|
```ini
|
||||||
|
7 1 * * * /scripts/backup/backup.sh -l /var/log/mailcow_backup.log > /dev/null 2>&1
|
||||||
Repo in directory *'MailcowBackup'*, all fields including pointers to additional
|
|
||||||
files to backup, exclusion patterns and a remote borg path. Prune: keep all
|
|
||||||
backups made in the last 14 days.
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
/var/borgbackup
|
|
||||||
/var/borgbackup/SSHprivate.key
|
|
||||||
myuser@usw-s001.rsync.net:MailcowBackup/
|
|
||||||
myPaSsWoRd
|
|
||||||
/root/scripts/xtraLocations.borg
|
|
||||||
/root/scripts/excludeLocations.borg
|
|
||||||
--keep-within=14d
|
|
||||||
borg1
|
|
||||||
```
|
```
|
||||||
|
|
||||||
Repo in directory *'myBackup'*, no exclusions, keep 14 days end-of-day, 52 weeks
|
## restoring backups
|
||||||
end-of-week
|
|
||||||
|
|
||||||
```Ini
|
Starting with version 3.0, a *restore.sh* file has been included to semi-automate restoring your backups to a clean mailcow instance. There are a few steps required and they are better explained in the wiki than would be possible in a short write-up like this. Please check out the [restore process overview](https://git.asifbacchus.app/asif/MailcowBackup/wiki/8.0-Restore-overview) for more information.
|
||||||
/var/borgbackup
|
|
||||||
/root/keys/rsyncPrivate.key
|
|
||||||
myuser@usw-s001.rsync.net:myBackup/
|
|
||||||
PaSsWoRd
|
|
||||||
/var/borgbackup/include.list
|
|
||||||
|
|
||||||
--keep-daily=14 --keep-weekly=52
|
## final notes
|
||||||
borg1
|
|
||||||
```
|
|
||||||
|
|
||||||
Repo in directory *'backup'*, no extra file locations, no exclusions, no remote
|
I think that's everything. For detailed information, please review the [wiki](https://git.asifbacchus.app/asif/MailcowBackup/wiki/_pages). If I've forgotten to document something there, please let me know. I know the wiki is long but, I hate how much stuff for Linux and open-source programs/scripts in general are so poorly documented especially for newbies and I didn't want to make that same mistake.
|
||||||
borg installation. Keep last 30 backups.
|
|
||||||
|
|
||||||
```Ini
|
I don't script too often and I'm a horrible programmer, so if you see anything that can be/should be improved, please let me know by filing an issue or submit your changes via a pull request! I love learning new ways of doing things and getting feedback, so suggestions and comments are more than welcome.
|
||||||
/root/.borg
|
|
||||||
/root/.borg/private.key
|
|
||||||
username@server.tld:backup/
|
|
||||||
pAsSw0rD
|
|
||||||
|
|
||||||
|
If this has helped you out, then please visit my blog at [https://mytechiethoughts.com](https://mytechiethoughts.com) where I solve problems like this all the time on a shoe-string or zero budget. Thanks!
|
||||||
--keep-within=30d
|
|
||||||
|
|
||||||
```
|
|
||||||
|
|
||||||
**Notice that the blank lines are very important!**
|
|
||||||
|
|
||||||
## 503 functionality
|
|
||||||
|
|
||||||
This script includes an entire section dedicated to copying an html file to act
|
|
||||||
as an error 503 notification page. Error 503 is by definition "service
|
|
||||||
temporarily unavailable" which is exactly the case for your Mailcow server
|
|
||||||
during a backup since the mail-flow containers have been disabled.
|
|
||||||
|
|
||||||
The script copies whatever file is defined by the *'-5'* parameter (or the
|
|
||||||
default located at *'scriptpath/503.html'*) to whatever path is defined as the
|
|
||||||
'webroot' by the *'-w'* parameter. This means that if you omit the *'-w'*
|
|
||||||
parameter, the script will necessarily skip this entire process and just issue a
|
|
||||||
warning to let you know about it.
|
|
||||||
|
|
||||||
### Conditional forwarding by your webserver
|
|
||||||
|
|
||||||
The script copying the file to the webroot is the easy part. Your webserver has
|
|
||||||
to look for the presence of that file and generate a 503 error in order for the
|
|
||||||
magic to happen. To do that, you have to include an instruction to that effect
|
|
||||||
in your default server definition and/or your Mailcow virtual server definition
|
|
||||||
file depending on your setup.
|
|
||||||
|
|
||||||
#### NGINX
|
|
||||||
|
|
||||||
You can copy the following code into the relevant server definition(s) on an
|
|
||||||
NGINX server:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
server {
|
|
||||||
...
|
|
||||||
if (-f /usr/share/nginx/html/503.html) {
|
|
||||||
return 503;
|
|
||||||
}
|
|
||||||
...
|
|
||||||
error_page 503 @backup
|
|
||||||
location @backup {
|
|
||||||
root /usr/share/nginx/html;
|
|
||||||
rewrite ^(.*)$ /503.html break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
This tells NGINX that if it finds the file *'503.html'* at the path
|
|
||||||
*'/usr/share/nginx/html'* (webroot on reverse proxy) then return an error code
|
|
||||||
503. Upon encountering a 503 error, rewrite any url to *'domain.tld/503.html'*
|
|
||||||
and thus, display the custom 503 error page. On the other hand, if it can't
|
|
||||||
find 503.html at the path specified (i.e. the script has deleted it because the
|
|
||||||
backup is completed), then go about business as usual.
|
|
||||||
|
|
||||||
#### Apache
|
|
||||||
|
|
||||||
I don't use apache for anything, ever... so I'm not sure how exactly you'd do
|
|
||||||
this but I think you'd have to use something like:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
RewriteEngine On
|
|
||||||
RewriteCond %{ENV:REDIRECT_STATUS} !=503
|
|
||||||
RewriteCond "/var/www/503.html" -f
|
|
||||||
RewriteRule ^ - [R=503,L]
|
|
||||||
...
|
|
||||||
ErrorDocument 503 /503.html
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
Let me know if that works and I'll update this document accordingly. Like I
|
|
||||||
said, I don't use Apache so I can't really test it very easily.
|
|
||||||
|
|
||||||
#### Disabling 503 functionality altogether
|
|
||||||
|
|
||||||
If you don't want to use the 503 functionality for whatever reason and don't
|
|
||||||
want your log file junked up with warnings about it, then find the section of
|
|
||||||
the script file that starts with *'--- Begin 503 section ---'* and either
|
|
||||||
comment all the lines (put a *'#'* at the beginning of each line) or delete all
|
|
||||||
the lines until you get to *'--- End 503 section ---'*.
|
|
||||||
|
|
||||||
## Scheduling: Cron
|
|
||||||
|
|
||||||
After running this script at least once manually to test your settings, you
|
|
||||||
should schedule it to run automatically so things stay backed up. This is
|
|
||||||
easiest with a simple cron job.
|
|
||||||
|
|
||||||
1. Open root's crontab:
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
sudo crontab -e
|
|
||||||
```
|
|
||||||
|
|
||||||
2. Add your script command line and set the time. I'm assuming your script is
|
|
||||||
located at *'/root/scripts'*, all files are at their default locations and
|
|
||||||
you want to run your backup at 1:07am daily.
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
7 1 * * * /root/scripts/backup.sh -l /var/log/backup.log -w /usr/share/nginx/html > /dev/null 2>&1
|
|
||||||
```
|
|
||||||
|
|
||||||
The last part redirects all output to 'null' and forwards any errors to
|
|
||||||
'null' also. You don't need output because the script creates a wonderfully
|
|
||||||
detailed log file that you can review :-)
|
|
||||||
3. Save the file and exit.
|
|
||||||
4. Confirm by listing the root user's crontab:
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
sudo crontab -l
|
|
||||||
```
|
|
||||||
|
|
||||||
## The log file
|
|
||||||
|
|
||||||
The script creates a very detailed log file of all major operations along with
|
|
||||||
any errors and warnings. Everything is timestamped so you can see how long
|
|
||||||
things take and when any errors took place. The script includes debugging
|
|
||||||
notes such as where temp files are located, where it's looking for data, whether
|
|
||||||
it created/moved/copied files, etc. All major operations are tagged *'-- [INFO]
|
|
||||||
message here --'*. Similarly, warnings are tagged *'-- [WARNING] message here
|
|
||||||
(code: xxxx) --'* and errors are tagged *'-- [ERROR] message here (code: xxx)
|
|
||||||
--'*. Successful operations generate a *'-- [SUCCESS] message here --'* stamp.
|
|
||||||
|
|
||||||
Sections of the script are all colour-coded to make viewing it easier. This
|
|
||||||
means you should use something like *'cat backup.log | more'* or *'tail -n
|
|
||||||
numberOfLines backup.log'* to view the file since the ansi colour codes
|
|
||||||
would make it difficult to read in nano or vi.
|
|
||||||
|
|
||||||
This tagging makes it easy for you to set up a log screening program to make
|
|
||||||
keeping an eye on your backup results easier. If you plan on using Logwatch
|
|
||||||
(highly recommended, great program!) then I've done the work for you...
|
|
||||||
|
|
||||||
### Using Logwatch
|
|
||||||
|
|
||||||
Log-group, conf and service files are included so that you can easily setup
|
|
||||||
Logwatch to monitor the script's log file and report at your desired detail
|
|
||||||
level as follows:
|
|
||||||
|
|
||||||
1. 0: Summary of total successes, warnings & errors only
|
|
||||||
2. 1-4: Actual success, error & warning messages
|
|
||||||
3. 5: Same as above, but includes info messages
|
|
||||||
4. 6+: Dumps entire raw log file including debugging messages
|
|
||||||
|
|
||||||
A detailed breakdown of the files and all options are included in a separate
|
|
||||||
readme in the *'/etc/logwatch'* folder of this git archive.
|
|
||||||
|
|
||||||
If you don't really care how it works, you can just copy the files from this
|
|
||||||
archive to your *'/etc/logwatch'* directory. The directory structure is correct
|
|
||||||
for a default Logwatch install on Debian/Ubuntu. You will have to update the
|
|
||||||
log-group file to reflect the path to your script's log file.
|
|
||||||
|
|
||||||
### Remember to rotate your logs
|
|
||||||
|
|
||||||
The log file generated by this script is fairly detailed so it can grow quite
|
|
||||||
large over time. This is especially true if you are using verbose output from
|
|
||||||
borg for any troubleshooting or for compliance/auditing. I've included a sample
|
|
||||||
commented *logrotate config file* in this git archive at *'/etc/logrotate.d'*
|
|
||||||
which you can modify and drop into that same directory on your Debian/Ubuntu
|
|
||||||
system. If you are using another log rotating solution, then please remember to
|
|
||||||
configure it so that your log files don't get overwhelmingly large should you
|
|
||||||
need to parse them if something goes wrong with your backups.
|
|
||||||
|
|
||||||
## Final notes
|
|
||||||
|
|
||||||
I think that's everything. If I've forgotten to document something, please let
|
|
||||||
me know. I know this readme is long but, I hate how much stuff for linux and
|
|
||||||
open-source programs/scripts in general are so poorly documented especially for
|
|
||||||
newbies and I didn't want to make that same mistake.
|
|
||||||
|
|
||||||
I don't script too often and I'm a horrible programmer, so if you see anything
|
|
||||||
that can be/should be improved, please let me know or submit your changes! I
|
|
||||||
love learning new ways of doing things and getting feedback, so suggestions and
|
|
||||||
comments are more than welcome.
|
|
||||||
|
|
||||||
If this has helped you out, then please visit my blog at
|
|
||||||
[https://mytechiethoughts.com](https://mytechiethoughts.com) where I solve
|
|
||||||
problems like this all the time on a shoe-string or zero budget. Thanks!
|
|
||||||
|
@ -18,11 +18,8 @@
|
|||||||
|
|
||||||
<body>
|
<body>
|
||||||
<h1>Bad timing!</h1>
|
<h1>Bad timing!</h1>
|
||||||
<p>Seems you're trying to access me during my daily backup window. Don't
|
<p>Seems you're trying to access me during my daily backup window. Don't worry though, I should be up and running again very soon.</p>
|
||||||
worry though, I should be up and running again very soon.</p>
|
<p>My average backup window duration is pretty short and I'm quite busy during that time copying your super-important stuff to my secure hiding place so they stay safe in case anything ever happens to me!</p>
|
||||||
<p>My average backup window duration is around 10 minutes during which time
|
|
||||||
I'm quite busy copying your super-important emails to my ultra-secret hiding
|
|
||||||
place so they stay safe in case anything ever happens to me!</p>
|
|
||||||
<h3><em>I'm really sorry for the delay. Please try me again soon!</em></h3>
|
<h3><em>I'm really sorry for the delay. Please try me again soon!</em></h3>
|
||||||
</body>
|
</body>
|
||||||
|
|
67
backup/backup.details
Normal file
67
backup/backup.details
Normal file
@ -0,0 +1,67 @@
|
|||||||
|
#######
|
||||||
|
### backup script configuration details
|
||||||
|
###
|
||||||
|
### This file contains sensitive information, make sure you have protected
|
||||||
|
### it by restricting permissions!
|
||||||
|
### Run the following in the directory where this file is located:
|
||||||
|
### chown root:root ./backup.details
|
||||||
|
### chmod 600 ./backup.details
|
||||||
|
###
|
||||||
|
### Do NOT include any commands in this file as they WILL be executed!!!
|
||||||
|
#######
|
||||||
|
|
||||||
|
|
||||||
|
### borg details
|
||||||
|
# if you're unsure what to enter here, please consult the repo wiki and/or
|
||||||
|
# the borg documentation
|
||||||
|
|
||||||
|
# base configuration directory for borg, all borg parameters use this directory
|
||||||
|
# as their 'root'. I recommend setups with this being "/var/borgbackup", the
|
||||||
|
# default is "$HOME" or "~$USER" in that order. If you're unsure, try "$HOME"
|
||||||
|
borgBaseDir="/var/borgbackup"
|
||||||
|
|
||||||
|
# SSH port on which your borg server listens. To connect on a custom port,
|
||||||
|
# uncomment the line below and change the port number as required.
|
||||||
|
# If you leave this line commented, the default (port 22) will be used.
|
||||||
|
#borgSSHPort=22
|
||||||
|
|
||||||
|
# full path to the SSH key used to connect to your remote backup server
|
||||||
|
borgSSHKey="/var/borgbackup/private.key"
|
||||||
|
|
||||||
|
# connection string to access the borg repo on your remote backup server
|
||||||
|
# this is usually in the form user@servername.tld:repoName/
|
||||||
|
borgConnectRepo="jdoe123@borg.server.net:mailcow/"
|
||||||
|
|
||||||
|
# password to access repo
|
||||||
|
# this was set when the repo was initialized and, while optional, is HIGHLY
|
||||||
|
# recommended for security
|
||||||
|
borgRepoPassphrase="p@ssW0rd"
|
||||||
|
|
||||||
|
# keyfile to access repo
|
||||||
|
# FULL PATH where the associated keyfile for your repo is located -- relevant
|
||||||
|
# only if your repo requires a keyfile (i.e. 'keyfile' vs 'repokey') and if you
|
||||||
|
# are not using the default keyfile location
|
||||||
|
#borgKeyfileLocation="/var/borgbackup/.config/borg/keys/server_address__repo_name"
|
||||||
|
|
||||||
|
# additional files to backup
|
||||||
|
# by default, the script will only backup your mailcow program directory
|
||||||
|
# (e.g. /opt/mailcow-dockerized) and your mailcow data volumes. If you would
|
||||||
|
# like additional files included (perhaps certificates, system configuration,
|
||||||
|
# etc.) include those locations listed one item per line in the file below
|
||||||
|
# see repo wiki for more details
|
||||||
|
borgXtraListPath="/scripts/backup/xtraLocations.borg"
|
||||||
|
|
||||||
|
# OPTIONAL: path to file containing files/directories or 'patterns' to be
|
||||||
|
# excluded in a BORG RECOGNIZED format
|
||||||
|
# see repo wiki for more details or consult borg documentation
|
||||||
|
# leave blank for no exclusions.
|
||||||
|
borgExcludeListPath="/scripts/backup/excludeLocations.borg"
|
||||||
|
|
||||||
|
# parameters to determine how borg deletes aged backups
|
||||||
|
# more details in the repo wiki and/or borg documentation
|
||||||
|
# leave blank to skip pruning altogether -- NOT recommended!
|
||||||
|
borgPruneSettings="--keep-within=14d --keep-daily=30 --keep-weekly=12 --keep-monthly=12"
|
||||||
|
|
||||||
|
# location of borg instance on your remote backup server
|
||||||
|
# this is very often just "borg1"
|
||||||
|
borgRemote="borg1"
|
861
backup/backup.sh
Executable file
861
backup/backup.sh
Executable file
@ -0,0 +1,861 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
#######
|
||||||
|
### mailcow backup using borgbackup
|
||||||
|
### this assumes three things:
|
||||||
|
### 1. standard mailcow-dockerized setup as per the docs
|
||||||
|
### 2. using borg to perform backups to ssh-capable remote server
|
||||||
|
### 3. remote repo already set-up and configured
|
||||||
|
#######
|
||||||
|
|
||||||
|
### text formatting presents
|
||||||
|
if command -v tput >/dev/null; then
|
||||||
|
bold=$(tput bold)
|
||||||
|
cyan=$(tput setaf 6)
|
||||||
|
err=$(tput bold)$(tput setaf 1)
|
||||||
|
magenta=$(tput setaf 5)
|
||||||
|
norm=$(tput sgr0)
|
||||||
|
ok=$(tput setaf 2)
|
||||||
|
warn=$(tput bold)$(tput setaf 3)
|
||||||
|
width=$(tput cols)
|
||||||
|
yellow=$(tput setaf 3)
|
||||||
|
else
|
||||||
|
bold=""
|
||||||
|
cyan=""
|
||||||
|
err=""
|
||||||
|
magenta=""
|
||||||
|
norm=""
|
||||||
|
ok=""
|
||||||
|
warn=""
|
||||||
|
width=80
|
||||||
|
yellow=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
### trap
|
||||||
|
trap trapExit 1 2 3 6
|
||||||
|
|
||||||
|
### functions
|
||||||
|
|
||||||
|
# bad configuration value passed in details file
|
||||||
|
badDetails() {
|
||||||
|
if [ "$1" = "empty" ]; then
|
||||||
|
exitError 130 "details:${2} cannot be NULL (undefined)"
|
||||||
|
elif [ "$1" = "dne" ]; then
|
||||||
|
exitError 131 "details:${2} file or directory does not exist."
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# bad parameter passed to script
|
||||||
|
badParam() {
|
||||||
|
if [ "$1" = "dne" ]; then
|
||||||
|
printf "\n%sError: '%s %s'\n" "$err" "$2" "$3"
|
||||||
|
printf "file or directory does not exist.%s\n\n" "$norm"
|
||||||
|
exit 1
|
||||||
|
elif [ "$1" = "empty" ]; then
|
||||||
|
printf "\n%sError: '%s' cannot have a NULL (empty) value.\n" "$err" "$2"
|
||||||
|
printf "%sPlease use '--help' for assistance%s\n\n" "$cyan" "$norm"
|
||||||
|
exit 1
|
||||||
|
elif [ "$1" = "svc" ]; then
|
||||||
|
printf "\n%sError: '%s %s': Service does not exist!%s\n\n" \
|
||||||
|
"$err" "$2" "$3" "$norm"
|
||||||
|
exit 1
|
||||||
|
elif [ "$1" = "user" ]; then
|
||||||
|
printf "\n%sError: '%s %s': User does not exist!%s\n\n" \
|
||||||
|
"$err" "$2" "$3" "$norm"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# cleanup
|
||||||
|
cleanup() {
|
||||||
|
# cleanup 503 if copied
|
||||||
|
if [ "$err503Copied" -eq 1 ]; then
|
||||||
|
# shellcheck disable=SC2129
|
||||||
|
if ! rm -f "$webroot/$err503File" 2>>"$logFile"; then
|
||||||
|
printf "%s[%s] -- [WARNING] Could not remove 503 error page." \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf " Web interface will not function until this file is " \
|
||||||
|
>>"$logFile"
|
||||||
|
printf "removed --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [INFO] 503 error page removed --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
# cleanup SQL dump directory if created
|
||||||
|
if [ "$sqlDumpDirCreated" -eq 1 ]; then
|
||||||
|
if ! rm -rf "$sqlDumpDir" 2>>"$logFile"; then
|
||||||
|
printf "%s[%s] -- [WARNING] Could not remove temporary SQL-dump directory. Sorry for the mess. --%s\n" \
|
||||||
|
"$warn" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [INFO] Temporary SQL-dump directory removed successfully --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
# start docker containers (no harm if they are already running)
|
||||||
|
doDocker start postfix
|
||||||
|
if [ "$dockerResultState" = "true" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] POSTFIX container is running --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 102 'Could not start POSTFIX container.' 'final'
|
||||||
|
fi
|
||||||
|
doDocker start dovecot
|
||||||
|
if [ "$dockerResultState" = "true" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] DOVECOT container is running --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 102 'Could not start DOVECOT container.' 'final'
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
doDocker() {
|
||||||
|
containerName="$(docker ps -a --format '{{ .Names }}' --filter "name=${COMPOSE_PROJECT_NAME}_${2}-mailcow_1" --filter "name=${COMPOSE_PROJECT_NAME}-${2}-mailcow-1")"
|
||||||
|
|
||||||
|
# determine action to take
|
||||||
|
if [ "$1" = "stop" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] Stopping %s-mailcow container --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$2" "$norm" >>"$logFile"
|
||||||
|
${dockerCmd} -f "$mcDockerCompose" stop --timeout "$dockerStopTimeout" "$2-mailcow" 2>>"$logFile"
|
||||||
|
# set result vars
|
||||||
|
dockerResultState="$(docker inspect -f '{{ .State.Running }}' "$containerName")"
|
||||||
|
dockerResultExit="$(docker inspect -f '{{ .State.ExitCode }}' "$containerName")"
|
||||||
|
elif [ "$1" = "start" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] Starting %s-mailcow container --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$2" "$norm" >>"$logFile"
|
||||||
|
${dockerCmd} -f "$mcDockerCompose" start "$2-mailcow" 2>>"$logFile"
|
||||||
|
# set result vars
|
||||||
|
dockerResultState="$(docker inspect -f '{{ .State.Running }}' "$containerName")"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# call cleanup and then exit with error report
|
||||||
|
exitError() {
|
||||||
|
printf "%s[%s] -- [ERROR] %s: %s --%s\n" \
|
||||||
|
"$err" "$(stamp)" "$1" "$2" "$norm" >>"$logFile"
|
||||||
|
if [ ! "$3" = "final" ]; then cleanup; fi
|
||||||
|
# note script completion with error
|
||||||
|
printf "%s[%s] --- %s execution completed with error ---%s\n" \
|
||||||
|
"$err" "$(stamp)" "$scriptName" "$norm" >>"$logFile"
|
||||||
|
exit "$1"
|
||||||
|
}
|
||||||
|
|
||||||
|
# display script help information
|
||||||
|
scriptHelp() {
|
||||||
|
newline
|
||||||
|
printf "%sUsage: %s [parameters]%s\n\n" "$bold" "$scriptName" "$norm"
|
||||||
|
textblock "There are NO mandatory parameters. If a parameter is not supplied, its default value will be used. In
|
||||||
|
the case of a switch parameter, it will remain *deactivated* if NOT specified."
|
||||||
|
newline
|
||||||
|
textblock "Parameters are listed then followed by a description of their effect on the following line. Finally, if a default value exists, it will be listed on the next line in (parentheses)."
|
||||||
|
newline
|
||||||
|
textblock "${magenta}--- script related parameters ---${norm}"
|
||||||
|
newline
|
||||||
|
switchTextblock "-b | --borg"
|
||||||
|
textblock "FULL path to borg executable file if not in the default location."
|
||||||
|
defaultsTextblock "(/usr/bin/borg)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-c | --config | --details"
|
||||||
|
textblock "Path to the configuration key/value-pair file for this script."
|
||||||
|
defaultsTextblock "(scriptPath/scriptName.details)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-h | -? | --help"
|
||||||
|
textblock "This help screen"
|
||||||
|
newline
|
||||||
|
switchTextblock "-l | --log"
|
||||||
|
textblock "Path to write log file"
|
||||||
|
defaultsTextblock "(scriptPath/scriptName.log)"
|
||||||
|
newline
|
||||||
|
switchTextblock "[SWITCH] --nc | --no-color | --no-colour"
|
||||||
|
textblock "Do NOT use ANSI colourization in the log file (in case your preferred log viewer does not support it)."
|
||||||
|
defaultsTextblock "(use ANSI colourization to make log file look good)"
|
||||||
|
newline
|
||||||
|
switchTextblock "--compression"
|
||||||
|
textblock "Compression algorithm(s) that borg should use. Please run 'borg help compression' for details."
|
||||||
|
defaultsTextblock "(not specified, use borg default of lz4)"
|
||||||
|
newline
|
||||||
|
switchTextblock "[SWITCH] -v | --verbose"
|
||||||
|
textblock "Log borg output with increased verbosity (list all files). Careful! Your log file can get very large very quickly!"
|
||||||
|
defaultsTextblock "(normal output, option is OFF)"
|
||||||
|
newline
|
||||||
|
textblock "${magenta}--- 503 functionality ---${norm}"
|
||||||
|
newline
|
||||||
|
switchTextblock "[SWITCH] -5 | --use-503"
|
||||||
|
textblock "Copy an 'error 503' page/indicator file to your webroot for your webserver to find. Specifying this option will enable other 503 options."
|
||||||
|
defaultsTextblock "(do NOT copy, option is OFF)"
|
||||||
|
newline
|
||||||
|
switchTextblock "--503-path"
|
||||||
|
textblock "Path to the file you want copied to your webroot as the 'error 503' page."
|
||||||
|
defaultsTextblock "(scriptPath/503_backup.html)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-w | --webroot"
|
||||||
|
textblock "Path to where the 'error 503' file should be copied."
|
||||||
|
defaultsTextblock "(/usr/share/nginx/html/)"
|
||||||
|
newline
|
||||||
|
textblock "${magenta}--- mailcow related ---${norm}"
|
||||||
|
newline
|
||||||
|
switchTextblock "-d | --docker-compose"
|
||||||
|
textblock "Path to the mailcow 'docker-compose.yml' file."
|
||||||
|
defaultsTextblock "(/opt/mailcow-dockerized/docker-compose.yml)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-m | --mailcow-config"
|
||||||
|
textblock "Path to mailcow configuration file ('mailcow.conf')."
|
||||||
|
defaultsTextblock "(/opt/mailcow-dockerized/mailcow.conf)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-t1 | --dockerStartTimeout"
|
||||||
|
textblock "Seconds to wait for docker containers to start."
|
||||||
|
defaultsTextblock "(180)"
|
||||||
|
newline
|
||||||
|
switchTextblock "-t2 | --dockerStopTimeout"
|
||||||
|
textblock "Seconds to wait for docker containers to stop."
|
||||||
|
defaultsTextblock "(120)"
|
||||||
|
newline
|
||||||
|
textblock "More details and examples of script usage can be found in the repo wiki at ${yellow}https://git.asifbacchus.app/asif/MailcowBackup/wiki${norm}"
|
||||||
|
newline
|
||||||
|
}
|
||||||
|
|
||||||
|
# generate dynamic timestamps
|
||||||
|
stamp() {
|
||||||
|
(date +%F" "%T)
|
||||||
|
}
|
||||||
|
|
||||||
|
textblock() {
|
||||||
|
printf "%s\n" "$1" | fold -w "$width" -s
|
||||||
|
}
|
||||||
|
|
||||||
|
defaultsTextblock() {
|
||||||
|
printf "%s%s%s\n" "$yellow" "$1" "$norm"
|
||||||
|
}
|
||||||
|
|
||||||
|
switchTextblock() {
|
||||||
|
printf "%s%s%s\n" "$cyan" "$1" "$norm"
|
||||||
|
}
|
||||||
|
|
||||||
|
# print a blank line
|
||||||
|
newline() {
|
||||||
|
printf "\n"
|
||||||
|
}
|
||||||
|
|
||||||
|
# same as exitError but for signal captures
|
||||||
|
trapExit() {
|
||||||
|
printf "%s[%s] -- [ERROR] 99: Caught signal --%s\n" \
|
||||||
|
"$err" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
cleanup
|
||||||
|
# note script completion with error
|
||||||
|
printf "%s[%s] --- %s execution was terminated via signal ---%s\n" \
|
||||||
|
"$err" "$(stamp)" "$scriptName" "$norm" >>"$logFile"
|
||||||
|
exit 99
|
||||||
|
}
|
||||||
|
|
||||||
|
### end of functions
|
||||||
|
|
||||||
|
### default variable values
|
||||||
|
|
||||||
|
## script related
|
||||||
|
# store logfile in the same directory as this script file using the same file
|
||||||
|
# name as the script but with the extension '.log'
|
||||||
|
scriptVersion="4.0"
|
||||||
|
scriptPath="$(CDPATH='' \cd -- "$(dirname -- "$0")" && pwd -P)"
|
||||||
|
scriptName="$(basename "$0")"
|
||||||
|
logFile="$scriptPath/${scriptName%.*}.log"
|
||||||
|
colourizeLogFile=1
|
||||||
|
warnCount=0
|
||||||
|
configDetails="$scriptPath/${scriptName%.*}.details"
|
||||||
|
err503Copied=0
|
||||||
|
sqlDumpDirCreated=0
|
||||||
|
exclusions=0
|
||||||
|
dockerCmd="docker compose"
|
||||||
|
borgPath="/usr/bin/borg"
|
||||||
|
# borg output verbosity -- normal
|
||||||
|
borgCreateParams='--stats'
|
||||||
|
borgPruneParams='--list'
|
||||||
|
|
||||||
|
# 503 related
|
||||||
|
use503=0
|
||||||
|
err503Path="$scriptPath/503_backup.html"
|
||||||
|
err503File="${err503Path##*/}"
|
||||||
|
webroot="/usr/share/nginx/html"
|
||||||
|
|
||||||
|
# mailcow/docker related
|
||||||
|
mcConfig='/opt/mailcow-dockerized/mailcow.conf'
|
||||||
|
mcDockerCompose="/opt/mailcow-dockerized/docker-compose.yml"
|
||||||
|
dockerStartTimeout=180
|
||||||
|
dockerStopTimeout=120
|
||||||
|
|
||||||
|
### process startup parameters
|
||||||
|
while [ $# -gt 0 ]; do
|
||||||
|
case "$1" in
|
||||||
|
-h | -\? | --help)
|
||||||
|
# display help
|
||||||
|
scriptHelp
|
||||||
|
exit 0
|
||||||
|
;;
|
||||||
|
--version)
|
||||||
|
# display script version
|
||||||
|
printf "\nMailcowBackup - Backup Mailcow using borgbackup to a remote SSH server : Version %s\n\n" ${scriptVersion}
|
||||||
|
exit 0
|
||||||
|
;;
|
||||||
|
-l | --log)
|
||||||
|
# set log file location
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
logFile="${2%/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
--nc | --no-color | --no-colour)
|
||||||
|
# do NOT colourize log file
|
||||||
|
colourizeLogFile=0
|
||||||
|
;;
|
||||||
|
-b | --borg)
|
||||||
|
# specify non-default borg path
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
borgPath="${2%/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-c | --config | --details)
|
||||||
|
# location of config details file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
configDetails="${2%/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam dne "$@"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
--compression)
|
||||||
|
# set borg archive compression
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
borgCompression="$2"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-v | --verbose)
|
||||||
|
# set verbose logging from borg
|
||||||
|
borgCreateParams='--list --stats'
|
||||||
|
borgPruneParams='--list'
|
||||||
|
;;
|
||||||
|
-5 | --use-503)
|
||||||
|
# enable copying 503 error page to webroot
|
||||||
|
use503=1
|
||||||
|
;;
|
||||||
|
--503-path)
|
||||||
|
# FULL path to 503 file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
err503Path="${2%/}"
|
||||||
|
err503File="${2##*/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam dne "$@"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-w | --webroot)
|
||||||
|
# path to webroot (copy 503)
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -d "$2" ]; then
|
||||||
|
webroot="${2%/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam dne "$@"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-d | --docker-compose)
|
||||||
|
# FULL path to mailcow docker-compose file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
mcDockerCompose="$2"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam dne "$@"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-m | --mailcow-config)
|
||||||
|
# FULL path to mailcow configuration file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
mcConfig="$2"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
badParam dne "$@"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
badParam empty "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-t1 | --timeout-start)
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
badParam empty "$@"
|
||||||
|
else
|
||||||
|
dockerStartTimeout="$2"
|
||||||
|
shift
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-t2 | --timeout-stop)
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
badParam empty "$@"
|
||||||
|
else
|
||||||
|
dockerStopTimeout="$2"
|
||||||
|
shift
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
printf "\n%sUnknown option: %s\n" "$err" "$1"
|
||||||
|
printf "%sUse '--help' for valid options.%s\n\n" "$cyan" "$norm"
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
shift
|
||||||
|
done
|
||||||
|
|
||||||
|
### check pre-requisites and default values
|
||||||
|
# check if running as root, otherwise exit
|
||||||
|
if [ "$(id -u)" -ne 0 ]; then
|
||||||
|
printf "\n%sERROR: script MUST be run as ROOT%s\n\n" "$err" "$norm"
|
||||||
|
exit 2
|
||||||
|
fi
|
||||||
|
# verify compose installed and set proper commands based on version
|
||||||
|
if ! (docker compose version >/dev/null 2>&1); then
|
||||||
|
if ! (docker-compose --version >/dev/null 2>&1); then
|
||||||
|
printf "\n%sERROR: Docker Compose not installed or not functioning%s\n\n" "$err" "$norm"
|
||||||
|
exit 3
|
||||||
|
fi
|
||||||
|
dockerCmd="docker-compose"
|
||||||
|
fi
|
||||||
|
# does the details file exist?
|
||||||
|
if [ ! -f "$configDetails" ]; then
|
||||||
|
badParam dne "(--details default)" "$configDetails"
|
||||||
|
fi
|
||||||
|
# is borg installed?
|
||||||
|
if ! find "$borgPath" -type f -executable >/dev/null 2>&1; then
|
||||||
|
printf "\n%sERROR: BORG cannot be found in the specified or default location on this system!%s\n\n" "$err" "$norm"
|
||||||
|
exit 4
|
||||||
|
fi
|
||||||
|
# if 503 functionality is enabled, do 503 related files exist?
|
||||||
|
if [ "$use503" -eq 1 ]; then
|
||||||
|
if [ ! -f "$err503Path" ]; then
|
||||||
|
badParam dne "(--503-path default)" "$err503Path"
|
||||||
|
elif [ ! -d "$webroot" ]; then
|
||||||
|
badParam dne "(--webroot default)" "$webroot"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
# verify mailcow.conf location and extract path
|
||||||
|
if [ ! -f "$mcConfig" ]; then
|
||||||
|
badParam dne "(--mailcow-config)" "$mcConfig"
|
||||||
|
fi
|
||||||
|
# verify docker-compose file exists
|
||||||
|
if [ ! -f "$mcDockerCompose" ]; then
|
||||||
|
badParam dne "(--docker-compose)" "$mcDockerCompose"
|
||||||
|
fi
|
||||||
|
# set compression level if specified
|
||||||
|
if [ -n "$borgCompression" ]; then
|
||||||
|
borgCreateParams="${borgCreateParams} --compression ${borgCompression}"
|
||||||
|
fi
|
||||||
|
# remove colourization if parameter specified
|
||||||
|
if [ "$colourizeLogFile" -eq 0 ]; then
|
||||||
|
bold=""
|
||||||
|
cyan=""
|
||||||
|
err=""
|
||||||
|
magenta=""
|
||||||
|
norm=""
|
||||||
|
ok=""
|
||||||
|
warn=""
|
||||||
|
yellow=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
### read mailcow.conf and set vars as needed
|
||||||
|
# shellcheck source=./mailcow.conf.shellcheck
|
||||||
|
. "$mcConfig"
|
||||||
|
export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
|
||||||
|
export COMPOSE_HTTP_TIMEOUT="$dockerStartTimeout"
|
||||||
|
|
||||||
|
### start logging
|
||||||
|
printf "%s[%s] --- Start %s execution ---%s\n" \
|
||||||
|
"$magenta" "$(stamp)" "$scriptName" "$norm" >>"$logFile"
|
||||||
|
printf "%s[%s] -- [INFO] Log located at %s%s%s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$yellow" "$logFile" "$cyan" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
### get location of docker volumes
|
||||||
|
dockerVolumeMail=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_vmail-vol-1)
|
||||||
|
printf "%s[%s] -- [INFO] Using MAIL volume: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$dockerVolumeMail" "$norm" >>"$logFile"
|
||||||
|
dockerVolumeRspamd=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_rspamd-vol-1)
|
||||||
|
printf "%s[%s] -- [INFO] Using RSPAMD volume: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$dockerVolumeRspamd" "$norm" >>"$logFile"
|
||||||
|
dockerVolumePostfix=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_postfix-vol-1)
|
||||||
|
printf "%s[%s] -- [INFO] Using POSTFIX volume: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$dockerVolumePostfix" "$norm" >>"$logFile"
|
||||||
|
dockerVolumeRedis=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_redis-vol-1)
|
||||||
|
printf "%s[%s] -- [INFO] Using REDIS volume: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$dockerVolumeRedis" "$norm" >>"$logFile"
|
||||||
|
dockerVolumeCrypt=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_crypt-vol-1)
|
||||||
|
printf "%s[%s] -- [INFO] Using MAILCRYPT volume: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$dockerVolumeCrypt" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
### read details file to get variables needed run borg
|
||||||
|
# check if config details file was provided as a relative or absolute path
|
||||||
|
case "${configDetails}" in
|
||||||
|
/*)
|
||||||
|
# absolute path, no need to rewrite variable
|
||||||
|
# shellcheck source=./backup.details
|
||||||
|
. "${configDetails}"
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
# relative path, prepend './' to create absolute path
|
||||||
|
# shellcheck source=./backup.details
|
||||||
|
. "./${configDetails}"
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
printf "%s[%s] -- [INFO] %s%s%s imported --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$yellow" "$configDetails" "$cyan" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
### Run borg variable checks
|
||||||
|
printf "%s[%s] -- [INFO] Verifying supplied borg details --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
## verify borg base directory
|
||||||
|
if [ -z "${borgBaseDir}" ]; then
|
||||||
|
badDetails empty 'borgBaseDir'
|
||||||
|
elif [ ! -d "${borgBaseDir}" ]; then
|
||||||
|
badDetails dne 'borgBaseDir'
|
||||||
|
fi
|
||||||
|
printf "%sdetails:borgBaseDir %s-- %s[OK]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
export BORG_BASE_DIR="${borgBaseDir%/}"
|
||||||
|
|
||||||
|
## check path to SSH keyfile
|
||||||
|
if [ -z "${borgSSHKey}" ]; then
|
||||||
|
badDetails empty 'borgSSHKey'
|
||||||
|
elif [ ! -f "${borgSSHKey}" ]; then
|
||||||
|
badDetails dne 'borgSSHKey'
|
||||||
|
fi
|
||||||
|
printf "%sdetails:borgSSHKey %s-- %s[OK]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
## check SSH port
|
||||||
|
if [ -z "${borgSSHPort}" ]; then
|
||||||
|
borgSSHPort=22
|
||||||
|
printf "%sdetails:borgSSHPort %s-- %s[DEFAULT]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
printf "%sdetails:borgSSHPort %s-- %s[CUSTOM]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
export BORG_RSH="ssh -i ${borgSSHKey} -p ${borgSSHPort}"
|
||||||
|
|
||||||
|
## check borg repo connect string
|
||||||
|
if [ -z "${borgConnectRepo}" ]; then
|
||||||
|
badDetails empty 'borgConnectRepo'
|
||||||
|
fi
|
||||||
|
printf "%sdetails:borgConnectRepo %s-- %s[OK]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
export BORG_REPO="${borgConnectRepo}"
|
||||||
|
|
||||||
|
## check borg repo password
|
||||||
|
if [ -n "${borgRepoPassphrase}" ]; then
|
||||||
|
printf "%sdetails:borgRepoPassphrase %s-- %s[OK]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
export BORG_PASSPHRASE="${borgRepoPassphrase}"
|
||||||
|
else
|
||||||
|
# if passwd is blank intentionally, this is insecure
|
||||||
|
printf "%s-- [WARNING] Using a borg repo without a password is an " \
|
||||||
|
"$warn" >>"$logFile"
|
||||||
|
printf "insecure configuration --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
# if this was an accident, we need to provide a bogus passwd so borg fails
|
||||||
|
# otherwise it will sit forever just waiting for input
|
||||||
|
export BORG_PASSPHRASE="DummyPasswordSoBorgFails"
|
||||||
|
fi
|
||||||
|
|
||||||
|
## check borg repository keyfile location
|
||||||
|
if [ -z "${borgKeyfileLocation}" ]; then
|
||||||
|
printf "%sdetails:borgKeyfileLocation %s-- %s[DEFAULT]%s\n" "$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
# check if keyfile location exists
|
||||||
|
if [ ! -f "${borgKeyfileLocation}" ]; then
|
||||||
|
badDetails dne 'borgKeyfileLocation'
|
||||||
|
fi
|
||||||
|
printf "%sdetails:borgKeyfileLocation %s-- %s[OK]%s\n" "$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
export BORG_KEY_FILE="${borgKeyfileLocation}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
## export borg remote path, if specified
|
||||||
|
if [ -n "${borgRemote}" ]; then export BORG_REMOTE_PATH="${borgRemote}"; fi
|
||||||
|
|
||||||
|
## check if exclusion list file is specified
|
||||||
|
if [ -n "${borgExcludeListPath}" ]; then
|
||||||
|
# check if the file actually exists
|
||||||
|
if [ ! -f "${borgExcludeListPath}" ]; then
|
||||||
|
badDetails dne 'borgExcludeListPath'
|
||||||
|
fi
|
||||||
|
exclusions=1
|
||||||
|
fi
|
||||||
|
|
||||||
|
## read additional files
|
||||||
|
if [ -z "${borgXtraListPath}" ]; then
|
||||||
|
xtraList=''
|
||||||
|
else
|
||||||
|
# check if file actually exists
|
||||||
|
if [ ! -f "${borgXtraListPath}" ]; then
|
||||||
|
badDetails dne 'borgXtraListPath'
|
||||||
|
fi
|
||||||
|
# read file contents into concatenated list for echo to cmdline
|
||||||
|
while read -r xtraItem; do
|
||||||
|
if [ -z "${xtraList}" ]; then
|
||||||
|
xtraList="${xtraItem}"
|
||||||
|
else
|
||||||
|
xtraList="${xtraList} ${xtraItem}"
|
||||||
|
fi
|
||||||
|
done <<EOF
|
||||||
|
$(sed -e '/^\s*#.*$/d' -e '/^\s*$/d' "${borgXtraListPath}")
|
||||||
|
EOF
|
||||||
|
printf "%sdetails:borgXtraListPath %s-- %s[OK]%s\n" \
|
||||||
|
"$magenta" "$norm" "$ok" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
|
||||||
|
### set location of sql dump
|
||||||
|
# this is done before resetting default TMP dir for borg
|
||||||
|
if ! sqlDumpDir=$(mktemp -d 2>/dev/null); then
|
||||||
|
exitError 115 'Unable to create temp directory for SQL dump.'
|
||||||
|
else
|
||||||
|
sqlDumpFile="backup-$(date +%Y%m%d_%H%M%S).sql"
|
||||||
|
sqlDumpDirCreated=1
|
||||||
|
printf "%s[%s] -- [INFO] SQL dump file will be stored at: %s --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$sqlDumpDir/$sqlDumpFile" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
|
||||||
|
### create borg temp dir:
|
||||||
|
## python requires a writable temporary directory when unpacking borg and
|
||||||
|
## executing commands. This defaults to /tmp but many systems mount /tmp with
|
||||||
|
## the 'noexec' option for security. Thus, we will use/create a 'tmp' folder
|
||||||
|
## within the BORG_BASE_DIR and instruct python to use that instead of /tmp
|
||||||
|
|
||||||
|
# check if BORG_BASE_DIR/tmp exists, if not, create it
|
||||||
|
if [ ! -d "${borgBaseDir}/tmp" ]; then
|
||||||
|
if ! mkdir "${borgBaseDir}/tmp"; then
|
||||||
|
exitError 132 "Unable to create borg ${borgBaseDir}/tmp directory"
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [INFO] Created %s%s/tmp " \
|
||||||
|
"$cyan" "$(stamp)" "$yellow" "${borgBaseDir}" >>"$logFile"
|
||||||
|
printf "%s--%s\n" "$cyan" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
export TMPDIR="${borgBaseDir}/tmp"
|
||||||
|
|
||||||
|
### 503 functionality
|
||||||
|
if [ "$use503" -eq 1 ]; then
|
||||||
|
printf "%s[%s] -- [INFO] Copying 503 error page to " \
|
||||||
|
"$cyan" "$(stamp)" >>"$logFile"
|
||||||
|
printf "webroot -- %s\n" "$norm" >>"$logFile"
|
||||||
|
if ! \cp --force "${err503Path}" "${webroot}/${err503File}" 2>>"$logFile"; then
|
||||||
|
printf "%s[%s] -- [WARNING] Failed to copy 503 error page. " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "Web users will NOT be notified --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [SUCCESS] 503 error page copied --%s\n" \
|
||||||
|
"$ok" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
# set cleanup flag
|
||||||
|
err503Copied=1
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
### change to mailcow directory so docker commands execute properly
|
||||||
|
\cd "${mcConfig%/*}" || exitError 100 'Could not change to mailcow directory.'
|
||||||
|
|
||||||
|
### stop postfix and dovecot mail containers to prevent mailflow during backup
|
||||||
|
doDocker stop postfix
|
||||||
|
if [ "$dockerResultState" = "false" ] && [ "$dockerResultExit" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [INFO] POSTFIX container stopped --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 101 'Could not stop POSTFIX container.'
|
||||||
|
fi
|
||||||
|
doDocker stop dovecot
|
||||||
|
if [ "$dockerResultState" = "false" ] && [ "$dockerResultExit" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [INFO] DOVECOT container stopped --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 101 'Could not stop DOVECOT container.'
|
||||||
|
fi
|
||||||
|
|
||||||
|
### dump SQL
|
||||||
|
printf "%s[%s] -- [INFO] Dumping mailcow SQL database --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
${dockerCmd} exec -T mysql-mailcow mysqldump --default-character-set=utf8mb4 \
|
||||||
|
-u${DBUSER} -p${DBPASS} ${DBNAME} >"$sqlDumpDir/$sqlDumpFile" 2>>"$logFile"
|
||||||
|
dumpResult=$(${dockerCmd} exec -T mysql-mailcow echo "$?")
|
||||||
|
if [ "$dumpResult" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [INFO] SQL database dumped successfully --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 118 'There was an error dumping the mailcow SQL database.'
|
||||||
|
fi
|
||||||
|
|
||||||
|
### dump redis inside container
|
||||||
|
# delete old redis dump if it exists
|
||||||
|
if [ -f "$dockerVolumeRedis/dump.rdb" ]; then
|
||||||
|
rm -f "$dockerVolumeRedis/dump.rdb"
|
||||||
|
fi
|
||||||
|
# dump redis
|
||||||
|
printf "%s[%s] -- [INFO] Dumping mailcow redis database --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
${dockerCmd} exec -T redis-mailcow redis-cli save >>"$logFile" 2>&1
|
||||||
|
rdumpResult=$(${dockerCmd} exec -T redis-mailcow echo "$?")
|
||||||
|
if [ "$rdumpResult" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [INFO] mailcow redis dumped successfully --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
exitError 119 'There was an error dumping the mailcow redis database.'
|
||||||
|
fi
|
||||||
|
|
||||||
|
### execute borg depending on whether exclusions are defined
|
||||||
|
printf "%s[%s] -- [INFO] Pre-backup tasks completed, calling borgbackup --%s\n" "$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
## construct the proper borg commandline
|
||||||
|
# base command
|
||||||
|
if [ "$exclusions" -eq 0 ]; then
|
||||||
|
borgCMD="${borgPath} create --show-rc ${borgCreateParams} \
|
||||||
|
::$(date +%Y-%m-%d_%H%M%S) \
|
||||||
|
${mcConfig%/*} \
|
||||||
|
${sqlDumpDir} \
|
||||||
|
${dockerVolumeMail} \
|
||||||
|
${dockerVolumeRspamd} \
|
||||||
|
${dockerVolumePostfix} \
|
||||||
|
${dockerVolumeRedis} \
|
||||||
|
${dockerVolumeCrypt} \
|
||||||
|
${xtraList}"
|
||||||
|
elif [ "$exclusions" -eq 1 ]; then
|
||||||
|
borgCMD="${borgPath} create --show-rc ${borgCreateParams} \
|
||||||
|
--exclude-from ${borgExcludeListPath} \
|
||||||
|
::$(date +%Y-%m-%d_%H%M%S) \
|
||||||
|
${mcConfig%/*} \
|
||||||
|
${sqlDumpDir} \
|
||||||
|
${dockerVolumeMail} \
|
||||||
|
${dockerVolumeRspamd} \
|
||||||
|
${dockerVolumePostfix} \
|
||||||
|
${dockerVolumeRedis} \
|
||||||
|
${dockerVolumeCrypt} \
|
||||||
|
${xtraList}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# execute borg
|
||||||
|
printf "%s[%s] -- [INFO] Executing borg backup operation --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
${borgCMD} 2>>"$logFile"
|
||||||
|
borgResult="$?"
|
||||||
|
|
||||||
|
## check borg exit status
|
||||||
|
if [ "$borgResult" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [SUCCESS] Borg backup completed --%s\n" \
|
||||||
|
"$ok" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
elif [ "$borgResult" -eq 1 ]; then
|
||||||
|
printf "%s[%s] -- [WARNING] Borg completed with warnings. " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "Review this logfile for details --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
elif [ "$borgResult" -ge 2 ]; then
|
||||||
|
err_1="Borg exited with a critical error. Please review this log file"
|
||||||
|
err_2="for details."
|
||||||
|
exitError 138 "$err_1 $err_2"
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [WARNING] Borg exited with unknown return code. " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "Review this logfile for details --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
fi
|
||||||
|
|
||||||
|
### execute borg prune if parameters are provided, otherwise skip with a warning
|
||||||
|
if [ -n "${borgPruneSettings}" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] Executing borg prune operation --%s\n" \
|
||||||
|
"$cyan" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
"${borgPath}" prune --show-rc -v ${borgPruneParams} ${borgPruneSettings} \
|
||||||
|
2>>"$logFile"
|
||||||
|
borgPruneResult="$?"
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [WARNING] No prune parameters provided. " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "Your archive will continue growing with each backup --%s\n" \
|
||||||
|
"$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
fi
|
||||||
|
|
||||||
|
## report on prune operation if executed
|
||||||
|
if [ -n "${borgPruneResult}" ]; then
|
||||||
|
if [ "${borgPruneResult}" -eq 0 ]; then
|
||||||
|
printf "%s[%s] -- [SUCCESS] Borg prune completed --%s\n" \
|
||||||
|
"$ok" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
elif [ "$borgPruneResult" -eq 1 ]; then
|
||||||
|
printf "%s[%s] -- [WARNING] Borg prune completed with warnings. " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "Review this logfile for details --%s\n" "$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
elif [ "$borgPruneResult" -ge 2 ]; then
|
||||||
|
err_1="Borg prune exited with a critical error. Please review this"
|
||||||
|
err_2="log file for details."
|
||||||
|
exitError 139 "$err_1 $err_2"
|
||||||
|
else
|
||||||
|
printf "%s[%s] -- [WARNING] Borg prune exited with an unknown " \
|
||||||
|
"$warn" "$(stamp)" >>"$logFile"
|
||||||
|
printf "return code. Review this logfile for details --%s\n" \
|
||||||
|
"$norm" >>"$logFile"
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
### all processes successfully completed, cleanup and exit gracefully
|
||||||
|
|
||||||
|
# note successful completion of borg commands
|
||||||
|
printf "%s[%s] -- [SUCCESS] Backup operations completed --%s\n" \
|
||||||
|
"$ok" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
|
||||||
|
# cleanup
|
||||||
|
cleanup
|
||||||
|
|
||||||
|
# note complete success, tally warnings and exit
|
||||||
|
printf "%s[%s] -- [SUCCESS] All processes completed --%s\n" \
|
||||||
|
"$ok" "$(stamp)" "$norm" >>"$logFile"
|
||||||
|
printf "%s[%s] --- %s execution completed ---%s\n" \
|
||||||
|
"$magenta" "$(stamp)" "$scriptName" "$norm" >>"$logFile"
|
||||||
|
if [ "$warnCount" -gt 0 ]; then
|
||||||
|
printf "%s%s warnings issued!%s\n" "$warn" "${warnCount}" "$norm" >>"$logFile"
|
||||||
|
else
|
||||||
|
printf "%s0 warnings issued.%s\n" "$ok" "$norm" >>"$logFile"
|
||||||
|
fi
|
||||||
|
exit 0
|
||||||
|
|
||||||
|
### error codes
|
||||||
|
# 1: parameter error
|
||||||
|
# 2: not run as root
|
||||||
|
# 3: docker compose not installed or not functioning
|
||||||
|
# 4: borg not installed
|
||||||
|
# 99: TERM signal trapped
|
||||||
|
# 100: could not change to mailcow-dockerized directory
|
||||||
|
# 101: could not stop container(s)
|
||||||
|
# 102: could not start container(s)
|
||||||
|
# 115: unable to create temp dir for SQL dump
|
||||||
|
# 118: error dumping SQL database
|
||||||
|
# 119: error dumping redis database
|
||||||
|
# 130: null configuration variable in details file
|
||||||
|
# 131: invalid configuration variable in details file
|
||||||
|
# 138: borg exited with a critical error
|
||||||
|
# 139: borg prune exited with a critical error
|
2
backup/excludeLocations.borg
Normal file
2
backup/excludeLocations.borg
Normal file
@ -0,0 +1,2 @@
|
|||||||
|
|
||||||
|
EOF
|
66
backup/mailcow.conf.shellcheck
Normal file
66
backup/mailcow.conf.shellcheck
Normal file
@ -0,0 +1,66 @@
|
|||||||
|
#
|
||||||
|
# sample mailcow configuration file
|
||||||
|
# (used by shellcheck)
|
||||||
|
#
|
||||||
|
|
||||||
|
MAILCOW_HOSTNAME=mail.example.com
|
||||||
|
|
||||||
|
DBNAME=mailcow
|
||||||
|
DBUSER=mailcow
|
||||||
|
DBPASS=verylongpassword
|
||||||
|
DBROOT=anotherlongpassword
|
||||||
|
|
||||||
|
HTTP_PORT=8080
|
||||||
|
HTTP_BIND=127.0.0.1
|
||||||
|
HTTPS_PORT=8443
|
||||||
|
HTTPS_BIND=127.0.0.1
|
||||||
|
|
||||||
|
SMTP_PORT=25
|
||||||
|
SMTPS_PORT=465
|
||||||
|
SUBMISSION_PORT=587
|
||||||
|
IMAP_PORT=143
|
||||||
|
IMAPS_PORT=993
|
||||||
|
POP_PORT=110
|
||||||
|
POPS_PORT=995
|
||||||
|
SIEVE_PORT=4190
|
||||||
|
DOVEADM_PORT=127.0.0.1:19991
|
||||||
|
SQL_PORT=127.0.0.1:13306
|
||||||
|
SOLR_PORT=127.0.0.1:18983
|
||||||
|
REDIS_PORT=127.0.0.1:7654
|
||||||
|
|
||||||
|
TZ=America/Edmonton
|
||||||
|
|
||||||
|
COMPOSE_PROJECT_NAME=mailcowdockerized
|
||||||
|
|
||||||
|
ACL_ANYONE=disallow
|
||||||
|
MAILDIR_GC_TIME=1440
|
||||||
|
ADDITIONAL_SAN=
|
||||||
|
SKIP_LETS_ENCRYPT=y
|
||||||
|
ENABLE_SSL_SNI=n
|
||||||
|
SKIP_IP_CHECK=n
|
||||||
|
SKIP_HTTP_VERIFICATION=n
|
||||||
|
SKIP_CLAMD=n
|
||||||
|
SKIP_SOGO=n
|
||||||
|
SKIP_SOLR=n
|
||||||
|
SOLR_HEAP=1024
|
||||||
|
USE_WATCHDOG=y
|
||||||
|
ALLOW_ADMIN_EMAIL_LOGIN=n
|
||||||
|
WATCHDOG_NOTIFY_EMAIL=a@example.com,b@example.com,c@example.com
|
||||||
|
WATCHDOG_NOTIFY_BAN=y
|
||||||
|
WATCHDOG_EXTERNAL_CHECKS=y
|
||||||
|
LOG_LINES=9999
|
||||||
|
|
||||||
|
IPV4_NETWORK=172.22.1
|
||||||
|
IPV6_NETWORK=fd4d:6169:6c63:6f77::/64
|
||||||
|
#SNAT_TO_SOURCE=
|
||||||
|
#SNAT6_TO_SOURCE=
|
||||||
|
|
||||||
|
#API_KEY=
|
||||||
|
#API_KEY_READ_ONLY=
|
||||||
|
#API_ALLOW_FROM=172.22.1.1,127.0.0.1
|
||||||
|
|
||||||
|
MAILDIR_SUB=Maildir
|
||||||
|
SOGO_EXPIRE_SESSION=480
|
||||||
|
DOVECOT_MASTER_USER=
|
||||||
|
DOVECOT_MASTER_PASS=
|
||||||
|
MAILCOW_PASS_SCHEME=BLF-CRYPT
|
666
backup/restore.sh
Executable file
666
backup/restore.sh
Executable file
@ -0,0 +1,666 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
#######
|
||||||
|
### restore mailcow from borgbackup repository
|
||||||
|
### this assumes three things:
|
||||||
|
### 1. standard mailcow-dockerized setup as per the docs
|
||||||
|
### 2. backups made using the backup script from this git repo
|
||||||
|
### 3. backups already downloaded from your borg repo
|
||||||
|
#######
|
||||||
|
|
||||||
|
### text-formatting presets
|
||||||
|
if command -v tput >/dev/null; then
|
||||||
|
bold=$(tput bold)
|
||||||
|
cyan=$(tput bold)$(tput setaf 6)
|
||||||
|
err=$(tput bold)$(tput setaf 1)
|
||||||
|
info=$(tput sgr0)
|
||||||
|
magenta=$(tput sgr0)$(tput setaf 5)
|
||||||
|
norm=$(tput sgr0)
|
||||||
|
ok=$(tput setaf 2)
|
||||||
|
width=$(tput cols)
|
||||||
|
yellow=$(tput sgr0)$(tput setaf 3)
|
||||||
|
else
|
||||||
|
bold=''
|
||||||
|
cyan=''
|
||||||
|
err=''
|
||||||
|
info=''
|
||||||
|
magenta=''
|
||||||
|
norm=''
|
||||||
|
ok=''
|
||||||
|
width=80
|
||||||
|
yellow=''
|
||||||
|
fi
|
||||||
|
|
||||||
|
### trap
|
||||||
|
trap trapExit 1 2 3 6
|
||||||
|
|
||||||
|
### functions
|
||||||
|
|
||||||
|
consoleError() {
|
||||||
|
printf "\n%s%s\n" "$err" "$2"
|
||||||
|
printf "Exiting.\n\n%s" "$norm"
|
||||||
|
exit "$1"
|
||||||
|
}
|
||||||
|
|
||||||
|
exitError() {
|
||||||
|
printf "%s[%s] --- %s execution completed with error ---\n%s" "$err" "$(stamp)" "$scriptName" "$norm" >>"$logfile"
|
||||||
|
exit "$1"
|
||||||
|
}
|
||||||
|
|
||||||
|
doRestore() {
|
||||||
|
sourceFiles=$(find "${backupLocation}" -iname "${1}" -type d)
|
||||||
|
if [ -n "$sourceFiles" ]; then
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
if ( ! (cd "$sourceFiles/_data" && tar -cf - .) | (cd "${2}" && tar xvf -) >>"$logfile"); then
|
||||||
|
return 1
|
||||||
|
else
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
if ( ! (cd "$sourceFiles/_data" && tar -cf - .) | (cd "${2}" && tar xvf -) >/dev/null 2>&1); then
|
||||||
|
return 1
|
||||||
|
else
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
return 2
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
scriptHelp() {
|
||||||
|
textNewline
|
||||||
|
printf "%sUsage: %s [parameters]%s\n\n" "$bold" "$scriptName" "$norm"
|
||||||
|
textblock "The only required parameter is -b | --backup-location."
|
||||||
|
textblock "If a parameter is not supplied, its default value will be used."
|
||||||
|
textblock "Switch parameters will only be activated if specified."
|
||||||
|
textblockHeader "script parameters"
|
||||||
|
textblockParam "-b | --backup-location"
|
||||||
|
textblock "Directory containing extracted backup files from borg repo. REQUIRED."
|
||||||
|
textNewline
|
||||||
|
textblockParam "-l | --log" "scriptPath/scriptName.log"
|
||||||
|
textblock "Path to write log file. Best efforts will be made to create any specified paths."
|
||||||
|
textNewline
|
||||||
|
textblockParam "-v | --verbose" "false"
|
||||||
|
textblock "Enable verbose logging. This will list EVERY restored file possibly making your log file quite large! [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "--skip-mail" "false"
|
||||||
|
textblock "Skip restoring mail and encryption key. [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "--skip-sql" "false"
|
||||||
|
textblock "Skip restoring mailcow settings database. [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "--skip-postfix" "false"
|
||||||
|
textblock "Skip restoring postfix settings. [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "--skip-rspamd" "false"
|
||||||
|
textblock "Skip restoring Rspamd settings/configuration/history. [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "--skip-redis" "false"
|
||||||
|
textblock "Skip restoring redis database. [SWITCH]"
|
||||||
|
textNewline
|
||||||
|
textblockParam "-? | -h | --help"
|
||||||
|
textblock "Display this help screen."
|
||||||
|
textblockHeader "mailcow parameters"
|
||||||
|
textblockParam "-d | --docker-compose" "/opt/mailcow-dockerized/docker-compose.yml"
|
||||||
|
textblock "FULL path to mailcow's 'docker-compose.yml' file."
|
||||||
|
textNewline
|
||||||
|
textblockParam "-m | --mailcow-config" "/opt/mailcow-dockerized/mailcow.conf"
|
||||||
|
textblock "FULL path to mailcow configuration file ('mailcow.conf'). The path of this file is also used to determine your mailcow directory."
|
||||||
|
textblockHeader "docker parameters"
|
||||||
|
textblockParam "-t1 | --timeout-start" "180"
|
||||||
|
textblock "Seconds to wait for docker containers to start."
|
||||||
|
textNewline
|
||||||
|
textblockParam "-t2 | --timeout-stop" "120"
|
||||||
|
textblock "Seconds to wait for docker containers to stop."
|
||||||
|
textNewline
|
||||||
|
textblock "More details and examples of script usage can be found in the repo wiki at ${yellow}https://git.asifbacchus.app/asif/MailcowBackup/wiki${norm}"
|
||||||
|
textNewline
|
||||||
|
exit 0
|
||||||
|
}
|
||||||
|
|
||||||
|
stamp() {
|
||||||
|
(date +%F' '%T)
|
||||||
|
}
|
||||||
|
|
||||||
|
textblock() {
|
||||||
|
printf "%s\n" "$1" | fold -w "$width" -s
|
||||||
|
}
|
||||||
|
|
||||||
|
textblockHeader() {
|
||||||
|
printf "\n%s%s*** %s ***%s\n\n" "$bold" "$magenta" "$1" "$norm"
|
||||||
|
}
|
||||||
|
|
||||||
|
textblockParam() {
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
# no default
|
||||||
|
printf "%s%s%s\n" "$cyan" "$1" "$norm"
|
||||||
|
else
|
||||||
|
# default parameter provided
|
||||||
|
printf "%s%s %s(%s)%s\n" "$cyan" "$1" "$yellow" "$2" "$norm"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
textNewline() {
|
||||||
|
printf "\n"
|
||||||
|
}
|
||||||
|
|
||||||
|
trapExit() {
|
||||||
|
printf "%s[%s] -- [ERROR] 99: Caught signal --%s\n" "$err" "$(stamp)" "$norm" >>"$logfile"
|
||||||
|
printf "%s[%s] --- %s execution terminated via signal ---\n%s" "$err" "$(stamp)" "$scriptName" "$norm" >>"$logfile"
|
||||||
|
exit 99
|
||||||
|
}
|
||||||
|
|
||||||
|
writeLog() {
|
||||||
|
if [ "$1" = "task" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] %s... " "$info" "$(stamp)" "$2" >>"$logfile"
|
||||||
|
elif [ "$1" = "done" ]; then
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
printf "%sdone%s --\n%s" "$ok" "$info" "$norm" >>"$logfile"
|
||||||
|
elif [ "$2" = "error" ]; then
|
||||||
|
printf "%sERROR%s --\n%s" "$err" "$info" "$norm" >>"$logfile"
|
||||||
|
elif [ "$2" = "warn" ]; then
|
||||||
|
printf "%swarning%s --\n%s" "$yellow" "$info" "$norm" >>"$logfile"
|
||||||
|
fi
|
||||||
|
elif [ "$1" = "error" ]; then
|
||||||
|
printf "%s[%s] -- [ERROR] %s: %s --\n%s" "$err" "$(stamp)" "$2" "$3" "$norm" >>"$logfile"
|
||||||
|
elif [ "$1" = "warn" ]; then
|
||||||
|
printf "%s[%s] -- [WARNING] %s --\n%s" "$yellow" "$(stamp)" "$2" "$norm" >>"$logfile"
|
||||||
|
elif [ "$1" = "info" ]; then
|
||||||
|
printf "%s[%s] -- [INFO] %s --\n%s" "$info" "$(stamp)" "$2" "$norm" >>"$logfile"
|
||||||
|
elif [ "$1" = "success" ]; then
|
||||||
|
printf "%s[%s] -- [SUCCESS] %s --\n%s" "$ok" "$(stamp)" "$2" "$norm" >>"$logfile"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
### parameter defaults
|
||||||
|
# script related
|
||||||
|
scriptVersion="4.0"
|
||||||
|
scriptPath="$(CDPATH='' cd -- "$(dirname -- "$0")" && pwd -P)"
|
||||||
|
scriptName="$(basename "$0")"
|
||||||
|
errorCount=0
|
||||||
|
warnCount=0
|
||||||
|
backupLocation=""
|
||||||
|
sqlBackup=""
|
||||||
|
restoreMail=1
|
||||||
|
restoreSQL=1
|
||||||
|
restorePostfix=1
|
||||||
|
restoreRedis=1
|
||||||
|
restoreRspamd=1
|
||||||
|
verbose=0
|
||||||
|
# logfile default: same location and name as script but with '.log' extension
|
||||||
|
logfile="$scriptPath/${scriptName%.*}.log"
|
||||||
|
# mailcow/docker related
|
||||||
|
mcConfig='/opt/mailcow-dockerized/mailcow.conf'
|
||||||
|
mcDockerCompose='/opt/mailcow-dockerized/docker-compose.yml'
|
||||||
|
sqlRunning=0
|
||||||
|
dockerStartTimeout=180
|
||||||
|
dockerStopTimeout=120
|
||||||
|
dockerCmd="docker compose"
|
||||||
|
|
||||||
|
### check if user is root
|
||||||
|
if [ "$(id -u)" -ne 0 ]; then
|
||||||
|
consoleError '2' "This script must be run as ROOT."
|
||||||
|
fi
|
||||||
|
|
||||||
|
### process startup parameters
|
||||||
|
while [ $# -gt 0 ]; do
|
||||||
|
case "$1" in
|
||||||
|
-h | -\? | --help)
|
||||||
|
# display help
|
||||||
|
scriptHelp
|
||||||
|
;;
|
||||||
|
--version)
|
||||||
|
# display script version
|
||||||
|
printf "\nMailcowRestore - Restore Mailcow from a borgbackup repo : Version %s\n\n" ${scriptVersion}
|
||||||
|
exit 0
|
||||||
|
;;
|
||||||
|
-l | --log)
|
||||||
|
# set logfile location
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
consoleError '1' "Log file path cannot be null. Leave unspecified to save log in the same directory as this script."
|
||||||
|
fi
|
||||||
|
logfile="$2"
|
||||||
|
shift
|
||||||
|
;;
|
||||||
|
-v | --verbose)
|
||||||
|
verbose=1
|
||||||
|
;;
|
||||||
|
-d | --docker-compose)
|
||||||
|
# FULL path to docker-compose file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
mcDockerCompose="$2"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot find docker-compose file as specified."
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot be blank/empty."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-m | --mailcow-config)
|
||||||
|
# FULL path to mailcow configuration file file
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -f "$2" ]; then
|
||||||
|
mcConfig="$2"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot find mailcow configuration file as specified."
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot be blank/empty."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-t1 | --timeout-start)
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
consoleError '1' "$1: cannot be blank/empty."
|
||||||
|
else
|
||||||
|
dockerStartTimeout="$2"
|
||||||
|
shift
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-t2 | --timeout-stop)
|
||||||
|
if [ -z "$2" ]; then
|
||||||
|
consoleError '1' "$1: cannot be blank/empty."
|
||||||
|
else
|
||||||
|
dockerStopTimeout="$2"
|
||||||
|
shift
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
-b | --backup-location)
|
||||||
|
if [ -n "$2" ]; then
|
||||||
|
if [ -d "$2" ] && [ -n "$(ls -A "$2")" ]; then
|
||||||
|
backupLocation="${2%/}"
|
||||||
|
shift
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot find specified backup location directory or it is empty."
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
consoleError '1' "$1: cannot be blank/empty."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
--skip-mail)
|
||||||
|
restoreMail=0
|
||||||
|
;;
|
||||||
|
--skip-sql)
|
||||||
|
restoreSQL=0
|
||||||
|
;;
|
||||||
|
--skip-postfix)
|
||||||
|
restorePostfix=0
|
||||||
|
;;
|
||||||
|
--skip-redis)
|
||||||
|
restoreRedis=0
|
||||||
|
;;
|
||||||
|
--skip-rspamd)
|
||||||
|
restoreRspamd=0
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
printf "\n%sUnknown option: %s\n" "$err" "$1"
|
||||||
|
printf "Use '--help' for valid options.%s\n\n" "$norm"
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
shift
|
||||||
|
done
|
||||||
|
|
||||||
|
### pre-flight checks
|
||||||
|
# ensure there's something to do
|
||||||
|
if [ "$restoreMail" -eq 0 ] && [ "$restoreSQL" -eq 0 ] && [ "$restorePostfix" -eq 0 ] && [ "$restoreRedis" -eq 0 ] && [ "$restoreRedis" -eq 0 ]; then
|
||||||
|
printf "\n%sAll restore operations skipped -- nothing to do!%s\n\n" "$yellow" "$norm"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
# set path so checks are valid for this script environment
|
||||||
|
export PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
|
||||||
|
# docker installed?
|
||||||
|
if ! command -v docker >/dev/null; then
|
||||||
|
consoleError '3' 'docker does not seem to be installed!'
|
||||||
|
fi
|
||||||
|
# verify compose installed and set proper commands based on version
|
||||||
|
if ! (docker compose version >/dev/null 2>&1); then
|
||||||
|
if ! (docker-compose --version >/dev/null 2>&1); then
|
||||||
|
printf "\n%sERROR: Docker Compose not installed or not functioning%s\n\n" "$err" "$norm"
|
||||||
|
exit 3
|
||||||
|
fi
|
||||||
|
dockerCmd="docker-compose"
|
||||||
|
fi
|
||||||
|
# mailcow.conf?
|
||||||
|
if [ ! -f "$mcConfig" ]; then
|
||||||
|
consoleError '1' "mailcow configuration file ($mcConfig) cannot be found."
|
||||||
|
fi
|
||||||
|
# docker-compose configuration?
|
||||||
|
if [ ! -f "$mcDockerCompose" ]; then
|
||||||
|
consoleError '1' "docker-compose configuration ($mcDockerCompose) cannot be found."
|
||||||
|
fi
|
||||||
|
# forgot to set backup location?
|
||||||
|
if [ -z "$backupLocation" ]; then
|
||||||
|
consoleError '1' "'--backup-location' cannot be unspecified or null/empty."
|
||||||
|
fi
|
||||||
|
# change to mailcow directory so commands execute properly
|
||||||
|
\cd "${mcConfig%/*}" || consoleError '4' 'Cannot change to mailcow directory as determined from mailcow.conf location.'
|
||||||
|
|
||||||
|
### read mailcow.conf and import vars
|
||||||
|
# shellcheck source=./mailcow.conf.shellcheck
|
||||||
|
. "$mcConfig"
|
||||||
|
export COMPOSE_HTTP_TIMEOUT="$dockerStartTimeout"
|
||||||
|
|
||||||
|
### start logging
|
||||||
|
# verify logfile specification is valid
|
||||||
|
if ! printf "%s" "$logfile" | grep -o / >/dev/null; then
|
||||||
|
# no slashes -> filename provided, save in scriptdir
|
||||||
|
logfile="$scriptPath/$logfile"
|
||||||
|
elif [ "$(printf "%s" "$logfile" | tail -c 1)" = '/' ]; then
|
||||||
|
# ends in '/' --> directory provided, does it exist?
|
||||||
|
if [ ! -d "$logfile" ]; then
|
||||||
|
if ! mkdir -p "$logfile" >/dev/null 2>&1; then
|
||||||
|
consoleError '1' "Unable to make specified log file directory."
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
logdir="$(cd "$logfile" 2>/dev/null && pwd -P)"
|
||||||
|
logfile="${logdir}/${scriptName%.*}.log"
|
||||||
|
else
|
||||||
|
# full path provided, does the parent directory exist?
|
||||||
|
if [ ! -d "${logfile%/*}" ]; then
|
||||||
|
# make parent path
|
||||||
|
if ! mkdir -p "${logfile%/*}" >/dev/null 2>&1; then
|
||||||
|
consoleError '1' "Unable to make specified log file path."
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
# write initial log entries
|
||||||
|
if ! printf "%s[%s] --- Start %s execution ---\n%s" "$magenta" "$(stamp)" "$scriptName" "$norm" 2>/dev/null >>"$logfile"; then
|
||||||
|
consoleError '1' "Unable to write to log file ($logfile)"
|
||||||
|
fi
|
||||||
|
writeLog 'info' "Log located at $logfile"
|
||||||
|
|
||||||
|
### get location of docker volumes
|
||||||
|
dockerVolumeMail=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_vmail-vol-1)
|
||||||
|
writeLog 'info' "Using MAIL volume: ${dockerVolumeMail}"
|
||||||
|
dockerVolumeCrypt=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_crypt-vol-1)
|
||||||
|
writeLog 'info' "Using MAILCRYPT volume: ${dockerVolumeCrypt}"
|
||||||
|
dockerVolumePostfix=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_postfix-vol-1)
|
||||||
|
writeLog 'info' "Using POSTFIX volume: ${dockerVolumePostfix}"
|
||||||
|
dockerVolumeRedis=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_redis-vol-1)
|
||||||
|
writeLog 'info' "Using REDIS volume: ${dockerVolumeRedis}"
|
||||||
|
dockerVolumeRspamd=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_rspamd-vol-1)
|
||||||
|
writeLog 'info' "Using RSPAMD volume: ${dockerVolumeRspamd}"
|
||||||
|
# exit if mail or crypt containers cannot be found (mailcow not initialized beforehand)
|
||||||
|
if [ -z "$dockerVolumeMail" ] || [ -z "$dockerVolumeCrypt" ]; then
|
||||||
|
writeLog 'error' '5' "Cannot find mail volume. Mailcow probably not initialized before running restore."
|
||||||
|
exitError 5
|
||||||
|
fi
|
||||||
|
|
||||||
|
### restore SQL
|
||||||
|
if [ "$restoreSQL" -eq 1 ]; then
|
||||||
|
writeLog 'task' "Restoring mailcow database"
|
||||||
|
|
||||||
|
# sql restore pre-requisites
|
||||||
|
sqlBackup=$(find "${backupLocation}/tmp" -iname "*.sql")
|
||||||
|
if [ -n "$sqlBackup" ]; then
|
||||||
|
# start mysql container if not already running
|
||||||
|
if ! docker container inspect -f '{{ .State.Running }}' ${COMPOSE_PROJECT_NAME}_mysql-mailcow_1 >/dev/null 2>&1; then
|
||||||
|
${dockerCmd} up -d mysql-mailcow >/dev/null 2>&1
|
||||||
|
if docker container inspect -f '{{ .State.Running }}' ${COMPOSE_PROJECT_NAME}_mysql-mailcow_1 >/dev/null 2>&1; then
|
||||||
|
sqlRunning=1
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '12' "Cannot start mysql-mailcow container -- cannot restore mailcow database!"
|
||||||
|
errorCount=$((errorCount + 1))
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
sqlRunning=1
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '11' "Cannot locate SQL backup -- cannot restore mailcow database!"
|
||||||
|
errorCount=$((errorCount + 1))
|
||||||
|
fi
|
||||||
|
|
||||||
|
# restore sql
|
||||||
|
if [ "$sqlRunning" -eq 1 ]; then
|
||||||
|
if docker exec -i "$(${dockerCmd} ps -q mysql-mailcow)" mysql -u${DBUSER} -p${DBPASS} ${DBNAME} <"${sqlBackup}" >/dev/null 2>&1; then
|
||||||
|
writeLog 'done'
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '13' "Something went wrong while trying to restore SQL database. Perhaps try again?"
|
||||||
|
errorCount=$((errorCount + 1))
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
### stop containers (necessary for all restore operations except SQL)
|
||||||
|
writeLog 'task' "Stopping mailcow"
|
||||||
|
if ! ${dockerCmd} down --timeout "${dockerStopTimeout}" >/dev/null 2>&1; then
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '20' "Unable to bring mailcow containers down -- cannot reliably restore. Aborting."
|
||||||
|
exitError 20
|
||||||
|
fi
|
||||||
|
if [ "$(docker ps --filter "name=${COMPOSE_PROJECT_NAME}" -q | wc -l)" -gt 0 ]; then
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '20' "Unable to bring mailcow containers down -- cannot reliably restore. Aborting."
|
||||||
|
exitError 20
|
||||||
|
fi
|
||||||
|
writeLog 'done'
|
||||||
|
|
||||||
|
### restore mail and encryption key
|
||||||
|
if [ "$restoreMail" -eq 1 ]; then
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'info' "Restoring email"
|
||||||
|
else
|
||||||
|
writeLog 'task' "Restoring email"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# restore email messages
|
||||||
|
doRestore "${COMPOSE_PROJECT_NAME}_vmail-vol-1" "$dockerVolumeMail"
|
||||||
|
ec="$?"
|
||||||
|
case "$ec" in
|
||||||
|
0)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'success' "Email messages restored"
|
||||||
|
else
|
||||||
|
writeLog 'done'
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
1)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more email messages."
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more email messages."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
2)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '51' "Cannot locate email message backups!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '51' "Cannot locate email message backups!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
# restore encryption key
|
||||||
|
doRestore "${COMPOSE_PROJECT_NAME}_crypt-vol-1" "$dockerVolumeCrypt"
|
||||||
|
ec="$?"
|
||||||
|
case "$ec" in
|
||||||
|
0)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'success' "Encryption key restored"
|
||||||
|
else
|
||||||
|
writeLog 'done'
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
1)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '52' "There was an error restoring the encryption key! Any restored messages are likely *not* readable!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '52' "There was an error restoring the encryption key! Any restored messages are likely *not* readable!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
2)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '51' "Cannot locate encryption key backup!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '51' "Cannot locate encryption key backup!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
fi
|
||||||
|
|
||||||
|
### restore postfix
|
||||||
|
if [ "$restorePostfix" -eq 1 ]; then
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'info' "Restoring postfix files"
|
||||||
|
else
|
||||||
|
writeLog 'task' "Restoring postfix files"
|
||||||
|
fi
|
||||||
|
|
||||||
|
doRestore "${COMPOSE_PROJECT_NAME}_postfix-vol-1" "$dockerVolumePostfix"
|
||||||
|
ec="$?"
|
||||||
|
case "$ec" in
|
||||||
|
0)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'success' "Postfix files restored"
|
||||||
|
else
|
||||||
|
writeLog 'done'
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
1)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more postfix files."
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more postfix files."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
2)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '51' "Cannot locate postfix backups!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '51' "Cannot locate postfix backups!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
fi
|
||||||
|
|
||||||
|
### restore rspamd
|
||||||
|
if [ "$restoreRspamd" -eq 1 ]; then
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'info' "Restoring Rspamd files"
|
||||||
|
else
|
||||||
|
writeLog 'task' "Restoring Rspamd files"
|
||||||
|
fi
|
||||||
|
|
||||||
|
doRestore "${COMPOSE_PROJECT_NAME}_rspamd-vol-1" "$dockerVolumeRspamd"
|
||||||
|
ec="$?"
|
||||||
|
case "$ec" in
|
||||||
|
0)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'success' "Rspamd files restored"
|
||||||
|
else
|
||||||
|
writeLog 'done'
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
1)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more Rspamd files."
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '52' "There was an error restoring one or more Rspamd files."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
2)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '51' "Cannot locate Rspamd backups!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '51' "Cannot locate Rspamd backups!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
fi
|
||||||
|
|
||||||
|
### restore redis
|
||||||
|
if [ "$restoreRedis" -eq 1 ]; then
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'info' "Restoring redis database"
|
||||||
|
else
|
||||||
|
writeLog 'task' "Restoring redis database"
|
||||||
|
fi
|
||||||
|
|
||||||
|
doRestore "${COMPOSE_PROJECT_NAME}_redis-vol-1" "$dockerVolumeRedis"
|
||||||
|
ec="$?"
|
||||||
|
case "$ec" in
|
||||||
|
0)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'success' "Redis database restored"
|
||||||
|
else
|
||||||
|
writeLog 'done'
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
1)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '52' "There was an error restoring the redis database. This is usually *not* a serious issue."
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '52' "There was an error restoring the redis database. This is usually *not* a serious issue."
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
2)
|
||||||
|
if [ "$verbose" -eq 1 ]; then
|
||||||
|
writeLog 'error' '51' "Cannot locate redis database backups!"
|
||||||
|
else
|
||||||
|
writeLog 'done' 'error'
|
||||||
|
writeLog 'error' '51' "Cannot locate redis database backups!"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
fi
|
||||||
|
|
||||||
|
### restart mailcow
|
||||||
|
writeLog 'task' "Starting mailcow"
|
||||||
|
if ! ${dockerCmd} up -d >/dev/null 2>&1; then
|
||||||
|
writeLog 'done' 'warn'
|
||||||
|
writeLog 'warn' '21' "Unable to automatically start mailcow containers. Please attempt a manual start and note any errors."
|
||||||
|
warnCount=$((warnCount + 1))
|
||||||
|
fi
|
||||||
|
writeLog 'done'
|
||||||
|
|
||||||
|
### exit gracefully
|
||||||
|
if [ "$errorCount" -gt 0 ]; then
|
||||||
|
# note non-terminating errors
|
||||||
|
printf "%s[%s] --- %s execution completed with %s error(s) ---\n%s" "$err" "$(stamp)" "$scriptName" "$errorCount" "$norm" >>"$logfile"
|
||||||
|
exit 98
|
||||||
|
elif [ "$warnCount" -gt 0 ]; then
|
||||||
|
printf "%s[%s] --- %s execution completed with %s warning(s) ---\n%s" "$yellow" "$(stamp)" "$scriptName" "$warnCount" "$norm" >>"$logfile"
|
||||||
|
exit 97
|
||||||
|
else
|
||||||
|
writeLog 'success' "All processes completed"
|
||||||
|
printf "%s[%s] --- %s execution completed ---\n%s" "$magenta" "$(stamp)" "$scriptName" "$norm" >>"$logfile"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
### error codes:
|
||||||
|
# 1: parameter error
|
||||||
|
# 2: not run as root
|
||||||
|
# 3: docker not installed
|
||||||
|
# 4: cannot change to mailcow directory
|
||||||
|
# 5: mailcow not initialized before running script
|
||||||
|
# 1x: SQL errors
|
||||||
|
# 11: cannot locate SQL dump in backup directory
|
||||||
|
# 12: cannot start mysql-mailcow container
|
||||||
|
# 13: restoring SQL dump was unsuccessful
|
||||||
|
# 2x: Docker/Docker-Compose errors
|
||||||
|
# 20: cannot bring docker container(s) down successfully
|
||||||
|
# 21: cannot bring docker container(s) up successfully
|
||||||
|
# 5x: File restore errors
|
||||||
|
# 51: cannot locate source files in backup directory
|
||||||
|
# 52: error restoring one or more files
|
||||||
|
# 97: script completed with 1 or more warnings
|
||||||
|
# 98: script completed with 1 or more non-terminating errors
|
||||||
|
# 99: TERM signal trapped
|
||||||
|
|
||||||
|
#EOF
|
60
backup/xtraLocations.borg
Normal file
60
backup/xtraLocations.borg
Normal file
@ -0,0 +1,60 @@
|
|||||||
|
# Files and directories listed here will be included in your borg backup
|
||||||
|
#
|
||||||
|
# Good candidates for inclusion would be things like your mailcow configuration
|
||||||
|
# files, customized docker-compose overrides, your SSL certificates, etc.
|
||||||
|
#
|
||||||
|
# List the path to files/directories one per line.
|
||||||
|
# Any blank lines will be ignored.
|
||||||
|
# Any lines starting with '#' will be ignored as a comment.
|
||||||
|
# For consistency, you should include the trailing slash for directories.
|
||||||
|
|
||||||
|
# these examples are for a very basic Debian machine hosting mailcow
|
||||||
|
|
||||||
|
|
||||||
|
### important system configuration files
|
||||||
|
|
||||||
|
# basic configuration
|
||||||
|
/etc/fstab
|
||||||
|
/etc/network/interfaces
|
||||||
|
/etc/network/interfaces.d/
|
||||||
|
/etc/systemd/timesyncd.conf
|
||||||
|
|
||||||
|
# ssh configuration and host keys
|
||||||
|
/etc/ssh/
|
||||||
|
|
||||||
|
# apt configuration
|
||||||
|
/etc/apt/sources.list
|
||||||
|
/etc/apt/sources.list.d/
|
||||||
|
/etc/apt/listchanges.conf
|
||||||
|
/etc/apt/apt.conf.d/50unattended-upgrades
|
||||||
|
/etc/apt/apt.conf.d/20auto-upgrades
|
||||||
|
|
||||||
|
# user profile defaults and configurations
|
||||||
|
/etc/profile
|
||||||
|
/etc/bash.bashrc
|
||||||
|
/etc/skel/
|
||||||
|
/etc/nanorc
|
||||||
|
|
||||||
|
# selected root user files
|
||||||
|
/root/.bashrc
|
||||||
|
/root/.ssh/
|
||||||
|
|
||||||
|
# scripts
|
||||||
|
/scripts/
|
||||||
|
|
||||||
|
|
||||||
|
### important programs and configurations
|
||||||
|
|
||||||
|
# name of program for reference
|
||||||
|
# include the paths to important configuration files/directories and/or
|
||||||
|
# data directories
|
||||||
|
|
||||||
|
# mailcow configuration
|
||||||
|
# already backed-up by the script by default based on mailcow.conf location
|
||||||
|
|
||||||
|
# NGINX (if this host is a reverse proxy, for example)
|
||||||
|
/etc/nginx/
|
||||||
|
/usr/share/nginx/html/
|
||||||
|
|
||||||
|
# LetsEncrypt
|
||||||
|
/etc/letsencrypt/
|
@ -1,300 +1,13 @@
|
|||||||
# Using Logwatch to monitor backup script <!-- omit in toc -->
|
# Using Logwatch to monitor the backup script
|
||||||
|
|
||||||
The backup script's log file has been set up so that utilities like Logwatch can
|
## quick start
|
||||||
easily parse it. In order to make that happen, a LogFile Group file, Service
|
|
||||||
and Script have to be created for Logwatch to generate reports. The correct
|
|
||||||
(general) directory structure has been created in this git archive already.
|
|
||||||
Below are the details of each file.
|
|
||||||
|
|
||||||
**If you don't care about how it works, you can simply copy this folder to your
|
Simply copy the contents of this folder to your logwatch configuration directory (*/etc/logwatch/* by default). The directory structure is already correct for a default Debian/Ubuntu logwatch installation. You **must** update the paths in */etc/logwatch/conf/logfiles/backup.conf* to point to your script's log file, but that's the only required change. Please consult [page 7.1.5](https://git.asifbacchus.app/asif/MailcowBackup/wiki/7.1.5-Testing) in the wiki for information on how to test logwatch using this new configuration.
|
||||||
Logwatch configuration directory (_/etc/logwatch/_ by default). Everything is
|
|
||||||
already in the proper directory structure for a default Debian/Ubuntu
|
|
||||||
installation.**
|
|
||||||
|
|
||||||
*If you need help installing or setting up Logwatch, please see my blog at
|
## more information
|
||||||
[https://mytechiethoughts.com](https://mytechiethoughts.com) and search for
|
|
||||||
'_logwatch_'*. These instructions assume you already have Logwatch setup correctly.
|
|
||||||
|
|
||||||
## Contents <!-- omit in toc -->
|
Please consult [section 7.1](https://git.asifbacchus.app/asif/MailcowBackup/wiki/7.1-Using-logwatch) in the wiki for detailed information about each logwatch configuration file contained within this section of the git repo and how to customize them for your environment.
|
||||||
|
|
||||||
- [LogFile Group file (/etc/logwatch/conf/logfiles/backup.conf)](#logfile-group-file-etclogwatchconflogfilesbackupconf)
|
## final thoughts
|
||||||
- [Log file location](#log-file-location)
|
|
||||||
- [Archive location and name format](#archive-location-and-name-format)
|
|
||||||
- [External script for timestamp processing](#external-script-for-timestamp-processing)
|
|
||||||
- [Service definition file (/etc/logwatch/conf/services/backup.conf)](#service-definition-file-etclogwatchconfservicesbackupconf)
|
|
||||||
- [LogFile Group file definition](#logfile-group-file-definition)
|
|
||||||
- [Report title](#report-title)
|
|
||||||
- [Detail level](#detail-level)
|
|
||||||
- [Service script (/etc/logwatch/scripts/services/backup)](#service-script-etclogwatchscriptsservicesbackup)
|
|
||||||
- [Detail levels](#detail-levels)
|
|
||||||
- [Timestamp processing script (/etc/logwatch/scripts/shared/sqfullstampanywhere)](#timestamp-processing-script-etclogwatchscriptssharedsqfullstampanywhere)
|
|
||||||
- [The time format specification](#the-time-format-specification)
|
|
||||||
- [The search REGEX](#the-search-regex)
|
|
||||||
- [Testing](#testing)
|
|
||||||
- [Final thoughts](#final-thoughts)
|
|
||||||
|
|
||||||
## LogFile Group file (/etc/logwatch/conf/logfiles/backup.conf)
|
I hope this helps you get your mailcow backup integrated with logwatch easily and quickly. If you have any suggestions/improvements, drop me a line in the issues section!
|
||||||
|
|
||||||
### Log file location
|
|
||||||
|
|
||||||
Update this as needed to point to the location and name of the log file
|
|
||||||
generated by the backup script. Remember, by default, the log file is created
|
|
||||||
in the same directory as the script itself.
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
LogFile = /path/to/your/backup.log
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
Best practices suggest you use the backup script's *-l* flag to change this
|
|
||||||
location to something like */var/log/backup.log*, for example. In that case,
|
|
||||||
the entry would look like:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
LogFile = /var/log/backup.log
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
### Archive location and name format
|
|
||||||
|
|
||||||
If you want Logwatch to process old (archived) log files generated by something
|
|
||||||
like *Logrotate*, then you have to specify the location and file name format of
|
|
||||||
those files. I've included the generalized compressed format of such rotated
|
|
||||||
files as the default in the script. Suppose you store your log files in the
|
|
||||||
recommended location (*/var/log/*) and are using *Logrotate* with compression
|
|
||||||
enabled, the archive line would look like:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
...
|
|
||||||
Archive = /var/log/backup.log.?.gz
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
This would tell Logwatch, when the archive option is set to true, that your
|
|
||||||
*backup.log* files are archived as: *backup.log.1.gz*, *backup.log.2.gz*, etc.
|
|
||||||
and are all located in */var/log/*.
|
|
||||||
|
|
||||||
**Note: This line is totally optional and only used if you set the archive
|
|
||||||
option in Logwatch to true (default). You can comment/delete this line if you
|
|
||||||
wish.**
|
|
||||||
|
|
||||||
### External script for timestamp processing
|
|
||||||
|
|
||||||
Since the log file uses a non-standard (according to Logwatch) method of
|
|
||||||
datestamping, a custom filter had to be created. See the
|
|
||||||
[relevant](#timestamp-processing-script-etclogwatchscriptssharedsqfullstampanywhere)
|
|
||||||
section of this document for more information.
|
|
||||||
|
|
||||||
The script file is called with an *\** before the filename.
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
...
|
|
||||||
*sqFullStampAnywhere
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
If you change the name of this file, you will have to change this line.
|
|
||||||
Remember that whatever you type here as a name is converted to all-lowercase
|
|
||||||
so your filename should be all lowercase also.
|
|
||||||
|
|
||||||
## Service definition file (/etc/logwatch/conf/services/backup.conf)
|
|
||||||
|
|
||||||
### LogFile Group file definition
|
|
||||||
|
|
||||||
The service file needs to know what group of log files it is responsible for
|
|
||||||
processing. This MUST match the name of your *LogFile Group file*:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
LogFile = backup
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
If you change your LogFile Group filename, then update it here too without the
|
|
||||||
*.conf* extension.
|
|
||||||
|
|
||||||
### Report title
|
|
||||||
|
|
||||||
The Logwatch output file (html or text) is divided into sections. You can
|
|
||||||
define the title to be anything that has meaning for you. I have arbitrarily
|
|
||||||
chosen *"System and Mailcow Backup"* but you can change it to anything you want by
|
|
||||||
modifying the line:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
...
|
|
||||||
Title = "System and Mailcow Backup"
|
|
||||||
```
|
|
||||||
|
|
||||||
### Detail level
|
|
||||||
|
|
||||||
If you want to set the *detail* level of this service differently from your
|
|
||||||
other services (which will use the *--detail* switch value or the value in your
|
|
||||||
*logwatch.conf*), then you can define that level here. By default, it appears
|
|
||||||
like this in the service configuration file:
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
...
|
|
||||||
# Override the detail level for this service
|
|
||||||
# Remember the levels are: 0, 1-4, 5, 6+
|
|
||||||
# Detail = 0
|
|
||||||
```
|
|
||||||
|
|
||||||
Simply change it to the value you want enforced. For example, here I'm setting
|
|
||||||
it to output level 5 regardless of whatever settings everything else is using.
|
|
||||||
|
|
||||||
```Ini
|
|
||||||
# Override the detail level for this service
|
|
||||||
# Remember the levels are: 0, 1-4, 5, 6+
|
|
||||||
Detail = 5
|
|
||||||
```
|
|
||||||
|
|
||||||
## Service script (/etc/logwatch/scripts/services/backup)
|
|
||||||
|
|
||||||
Logwatch calls any script with a name that **matches the service name**. You'll
|
|
||||||
notice that I just named everything *backup* to keep things simple. You can
|
|
||||||
change this to whatever you want, however. If you changed the service name to
|
|
||||||
*"MailcowBackup*.conf", for example, you would have to rename this script file
|
|
||||||
to "*MailcowBackup*" with no extension.\
|
|
||||||
*Note: The script is a PERL file (note the shebang) but it can be written in any
|
|
||||||
language.*
|
|
||||||
|
|
||||||
**In essence, Logwatch just spits out the log file(s) defined in the LogFile
|
|
||||||
Group file as standard input (STDIN) for the script and then takes whatever is
|
|
||||||
output (STDOUT) from the script to assemble into it's report.**
|
|
||||||
|
|
||||||
### Detail levels
|
|
||||||
|
|
||||||
The script supports four (4) detail levels as follows:
|
|
||||||
|
|
||||||
- **Level 0: Summary output only**
|
|
||||||
- This will display an aggregate total of certain logged elements. It will
|
|
||||||
display the total number of overall successful script executions, total
|
|
||||||
generated warnings and total errors encountered that stopped the normal
|
|
||||||
execution of the script. All totals are relative to the reporting period
|
|
||||||
Logwatch is using (--range parameter).
|
|
||||||
|
|
||||||
**This is the recommended reporting level.** It does not take up much space
|
|
||||||
and is quick to read. If you notice warnings and/or errors, you should
|
|
||||||
consult the full log.
|
|
||||||
- **Levels 1-4 (all the same): Critical messages**
|
|
||||||
- This uses the data which is summarized by Level 0 but outputs the actual
|
|
||||||
messages in the log file. For example, you will see the actual text of the
|
|
||||||
errors logged instead of just a total number of errors. This level of
|
|
||||||
reporting is useful when *initially* monitoring the script's operation since
|
|
||||||
you can see the actual text of any generated warnings/errors.
|
|
||||||
|
|
||||||
- **Level 5: Verbose (debugging) output**
|
|
||||||
- Like the previous level, this outputs the actual messages found in the log
|
|
||||||
file. However, it also includes *[INFO] tags* which contain logged
|
|
||||||
operational messages such as created temporary directories,
|
|
||||||
starting/stopping docker containers, whether the 503 page was copied, etc.
|
|
||||||
This level of reporting is useful in diagnosing why errors are occurring or
|
|
||||||
if you just want more insight into how the script works.
|
|
||||||
|
|
||||||
**This level of output will make your Logwatch reports longer and consume
|
|
||||||
more of your time to review. You should not use this level day-to-day.**
|
|
||||||
- **Levels 6+ (all the same): Complete log file dump**
|
|
||||||
- Any number greater than 5 passed as a detail level will trigger the script
|
|
||||||
to dump the entire log file out to Logwatch line-by-line. This is useful
|
|
||||||
only if you are debugging an issue and cannot get access to the actual raw
|
|
||||||
log file itself. The actual log file is colour-coded which makes it much
|
|
||||||
easier to read for debugging purposes.
|
|
||||||
|
|
||||||
**Use this detail level only when you need to see the entire log file and
|
|
||||||
cannot otherwise access the log file.**
|
|
||||||
|
|
||||||
## Timestamp processing script (/etc/logwatch/scripts/shared/sqfullstampanywhere)
|
|
||||||
|
|
||||||
This is basically a modified version of the '*applyeurodate*' script that comes
|
|
||||||
with Logwatch. It had to be modified to search within [square brackets] and to
|
|
||||||
accept characters coming before the stamp (i.e. ANSI colour codes). If you
|
|
||||||
change the '**stamp**' variable in the backup script to update the timestamp to
|
|
||||||
your liking (which to totally fine!) then you'll probably have to update this
|
|
||||||
file. There are two lines you need to modify to suit your new '**stamp**'
|
|
||||||
variable.
|
|
||||||
|
|
||||||
### The time format specification
|
|
||||||
|
|
||||||
'*$SearchDate*' is the variable used in the PERL script to do exactly what it
|
|
||||||
says, search for the date stamp. I have it set up to look for the format
|
|
||||||
'*year-month-date hour:minute:second*'. Note, we don't care about brackets or
|
|
||||||
anything here, we're just defining the format of the date/time stamp.
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
$SearchDate = TimeFilter('%Y-%m-%d %H:%M:%S');
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
If you changed the '**stamp**' variable so it was formatted as '*month/day/year
|
|
||||||
hour:minute*' (ex: '*[09/27/2018 18:38]*') then you'd update the **$SearchDate**
|
|
||||||
variable as follows (note: no mention of the square brackets!):
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
$SearchDate = TimeFilter('%m/%d/%Y %H:%M');
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
### The search REGEX
|
|
||||||
|
|
||||||
The PERL script uses a '*regular expression*' (REGEX) to search within the log file for
|
|
||||||
'*$SearchDate*'. For the default datestamp, this specification looks like:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
if ($ThisLine =~ m/\[$SearchDate\] /o) {
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
The REGEX appears between '*m/*' and '*/o*'. In this case, it searches for
|
|
||||||
'*$SearchDate*' inside [square brackets] appearing anywhere on the line. This
|
|
||||||
is because ANSI colour-codes often appear before the datestamp in the default
|
|
||||||
log file. If you have modified this so that your datestamp appears at the
|
|
||||||
beginning of the line and in the example format in the section above (using
|
|
||||||
slashes instead of dashes) then you'd rewrite this REGEX as follows:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
if ($ThisLine =~ m/^\[$SearchDate\] /o) {
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
or using regular brackets anywhere on the line:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
if ($ThisLine =~ m/\($SearchDate\) /o) {
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
or without any brackets but appearing at the beginning of the line:
|
|
||||||
|
|
||||||
```Perl
|
|
||||||
...
|
|
||||||
if ($ThisLine =~ m/^$SearchDate /o) {
|
|
||||||
...
|
|
||||||
```
|
|
||||||
|
|
||||||
## Testing
|
|
||||||
|
|
||||||
Run *logwatch --help* and note the options. You can test just this service
|
|
||||||
locally on your screen with the following command (assuming you kept default
|
|
||||||
names for everything):
|
|
||||||
|
|
||||||
```Bash
|
|
||||||
# Summary output, entire duration of log file
|
|
||||||
logwatch --service backup --output stdout --format text --range all --detail 0
|
|
||||||
|
|
||||||
# Minimal detail, yesterday only
|
|
||||||
logwatch --service backup --output stdout --format text --range yesterday --detail 3
|
|
||||||
|
|
||||||
# Verbose output, today only
|
|
||||||
logwatch --service backup --output stdout --format text --range today --detail 5
|
|
||||||
```
|
|
||||||
|
|
||||||
## Final thoughts
|
|
||||||
|
|
||||||
That's it! I'm a horrible PERL programmer so if anyone can optimize/improve the
|
|
||||||
script file used for Logwatch then please do it! Otherwise, I hope this made
|
|
||||||
sense and helped you integrate the backup script with Logwatch for easy
|
|
||||||
monitoring :-)
|
|
||||||
|
@ -3,8 +3,8 @@
|
|||||||
#############################################################################
|
#############################################################################
|
||||||
# $Id$
|
# $Id$
|
||||||
#############################################################################
|
#############################################################################
|
||||||
# Log: Backup script (backup)
|
# Log: mailcow backup
|
||||||
# Revision 1.0 2018/10/16
|
# Revision 1.1 2019/07/20
|
||||||
# Written by Asif Bacchus
|
# Written by Asif Bacchus
|
||||||
#############################################################################
|
#############################################################################
|
||||||
|
|
||||||
@ -34,7 +34,7 @@ if ($detailLevel == 0) {
|
|||||||
elsif ($ThisLine =~ /\-- \[WARNING\] /) {
|
elsif ($ThisLine =~ /\-- \[WARNING\] /) {
|
||||||
$summaryWarn++;
|
$summaryWarn++;
|
||||||
}
|
}
|
||||||
elsif ($ThisLine =~ /All processes completed successfully/) {
|
elsif ($ThisLine =~ /All processes completed/) {
|
||||||
$summarySuccess++;
|
$summarySuccess++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,881 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
|
|
||||||
### Text formatting presets
|
|
||||||
normal="\e[0m"
|
|
||||||
bold="\e[1m"
|
|
||||||
default="\e[39m"
|
|
||||||
err="\e[1;31m"
|
|
||||||
warn="\e[1;93m"
|
|
||||||
ok="\e[32m"
|
|
||||||
lit="\e[93m"
|
|
||||||
op="\e[39m"
|
|
||||||
info="\e[96m"
|
|
||||||
note="\e[95m"
|
|
||||||
|
|
||||||
|
|
||||||
### Functions ###
|
|
||||||
|
|
||||||
### scriptHelp -- display usage information for this script
|
|
||||||
function scriptHelp {
|
|
||||||
echo -e "${bold}${note}\n${scriptName} usage instructions:\n${normal}"
|
|
||||||
echo -e "${default}This script performs a backup of your Mailcow system"
|
|
||||||
echo -e "assuming a fairly standard set up such as outlined at"
|
|
||||||
echo -e "${lit}https://mytechiethoughts.com${default}. Full details about"
|
|
||||||
echo -e "this script can be found at that site."
|
|
||||||
echo -e "${bold}\nThe script performs the following tasks:${normal}${default}"
|
|
||||||
echo -e "1. Stops mailflow (postfix & dovecot) containers."
|
|
||||||
echo -e "2. Optionally copies a 503 error page to your webroot."
|
|
||||||
echo -e "3. Dumps Mailcow's SQL DB to a temporary directory."
|
|
||||||
echo -e "4. Invokes borgbackup to backup your SQL info, Mailcow settings"
|
|
||||||
echo -e "\tand (raw) data files along with any other files you specify."
|
|
||||||
echo -e "5. Removes temp files, the 503 error page and restores"
|
|
||||||
echo -e "\tMailcow to operational status."
|
|
||||||
echo -e "\nThe readme file included in this script's git contains detailed"
|
|
||||||
echo -e "usage information. The following is a brief summary:\n"
|
|
||||||
echo -e "${bold}***You MUST provide the full path to your mailcow"
|
|
||||||
echo -e "configuration file using the '-m' parameter***${normal}${default}"
|
|
||||||
echo -e "${bold}${note}\nMandatory parameters:${normal}${default}"
|
|
||||||
echo -e "${lit}\n-m, File name of the Mailcow build configuration file${default}"
|
|
||||||
echo -e "FULL PATH to your Mailcow master build configuration file containing"
|
|
||||||
echo -e "all variables and configuration info unique to your Mailcow setup."
|
|
||||||
echo -e "The path specified here is also used for all docker-related"
|
|
||||||
echo -e "operations in this script."
|
|
||||||
echo -e "${info}Default: <none>${default}"
|
|
||||||
echo -e "${bold}${note}\nOptional parameters:${normal}${default}"
|
|
||||||
echo -e "${lit}\n-1, Timeout for containers to STOP before error${default}"
|
|
||||||
echo -e "The number of seconds to wait for a docker container to STOP"
|
|
||||||
echo -e "before aborting the procedure and exiting this script with an"
|
|
||||||
echo -e "error."
|
|
||||||
echo -e "${info}Default: 120 seconds${default}"
|
|
||||||
echo -e "${lit}\n-2, Timeout for containers to START before error${default}"
|
|
||||||
echo -e "The number of seconds to wait for a docker container to START"
|
|
||||||
echo -e "before aborting the procedure and exiting this script with an"
|
|
||||||
echo -e "error."
|
|
||||||
echo -e "${info}Default: 180 seconds${default}"
|
|
||||||
echo -e "${lit}\n-5, Location of 503 error page file${default}"
|
|
||||||
echo -e "FULL PATH to the 503 error page HTML file you want copied to your"
|
|
||||||
echo -e "webroot to inform users the server is down during the backup. If"
|
|
||||||
echo -e "you don't specify a path/file, the default will be used. If the"
|
|
||||||
echo -e "default cannot be found, a warning will be logged and the script"
|
|
||||||
echo -e "will continue."
|
|
||||||
echo -e "${info}Default: ScriptPath/503.html${default}"
|
|
||||||
echo -e "${lit}\n-b, Location of file with borg repo details${default}"
|
|
||||||
echo -e "FULL PATH to the plain text file containing all information needed"
|
|
||||||
echo -e "to connect and process your borg repo. Details on the structure of"
|
|
||||||
echo -e "this file are in the readme and on ${lit}https://mytechiethoughts.com${default}"
|
|
||||||
echo -e "${info}Default: ScriptPath/nc_borg.details${default}"
|
|
||||||
echo -e "${lit}\n-d, File name of the docker-compose configuration file${default}"
|
|
||||||
echo -e "Name of the docker-compose configuration file that Mailcow uses"
|
|
||||||
echo -e "to build/start/stop all containers. This will only be searched"
|
|
||||||
echo -e "for in the path provided for your mailcow configuration file."
|
|
||||||
echo -e "${info}Default: docker-compose.yml${default}"
|
|
||||||
echo -e "${lit}\n-l, Location to save log file${default}"
|
|
||||||
echo -e "This script writes a detailed log file of all activities. It is"
|
|
||||||
echo -e "structured in an way easy for log parsers (like Logwatch) to read."
|
|
||||||
echo -e "${info}Default: ScriptPath/ScriptName.log${default}"
|
|
||||||
echo -e "${lit}\n-v, Verbose output from borgbackup${default}"
|
|
||||||
echo -e "By default, this script will only log summary data from borg."
|
|
||||||
echo -e "If you need/want more detailed information, the verbose setting"
|
|
||||||
echo -e "will list every file processed along with their status. Note: Your"
|
|
||||||
echo -e "log file can quickly get very very large using this option!"
|
|
||||||
echo -e "${info}Default: NOT activated (standard logging)${default}"
|
|
||||||
echo -e "${lit}\n-w, webserver's webroot directory${default}"
|
|
||||||
echo -e "This is the location from which your webserver (NGINX, Apache,"
|
|
||||||
echo -e "etc.) physically stores files to be served. This is NOT the"
|
|
||||||
echo -e "configuration directory for your webserver! It is the place"
|
|
||||||
echo -e "where the actual HTML/PHP/CSS/JS/etc. files are stored."
|
|
||||||
echo -e "NOTE: If you omit this option, then the entire 503 copy process"
|
|
||||||
echo -e "will be skipped regardless of the presence of a 503.html file."
|
|
||||||
echo -e "If you don't want to use the 503 feature, omitting this is an easy"
|
|
||||||
echo -e "way to skip it!"
|
|
||||||
echo -e "${info}Default: NONE${default}"
|
|
||||||
echo -e "${lit}\n-?, This help screen${default}\n"
|
|
||||||
echo -e "${bold}Please refer to the readme file and/or ${lit}https://mytechiethoughts.com${default}"
|
|
||||||
echo -e "for more information on this script.${normal}\n"
|
|
||||||
# exit with code 1 -- there is no use logging this
|
|
||||||
exit 1
|
|
||||||
}
|
|
||||||
|
|
||||||
### Generate dynamic timestamp
|
|
||||||
function stamp {
|
|
||||||
echo `date +%F" "%T`
|
|
||||||
}
|
|
||||||
|
|
||||||
### quit -- exit the script after logging any errors, warnings, etc.
|
|
||||||
function quit {
|
|
||||||
# list generated warnings, if any
|
|
||||||
if [ ${#exitWarn[@]} -gt 0 ]; then
|
|
||||||
echo -e "\n${warn}${scriptName} generated the following warnings:" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
for warnCode in "${exitWarn[@]}"; do
|
|
||||||
warnStamp="${warnCode%%_*}"
|
|
||||||
warnValue="${warnCode##*_}"
|
|
||||||
echo -e "${warn}${warnStamp} -- [WARNING]" \
|
|
||||||
"${warningExplain[$warnValue]} (code: ${warnValue}) --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
done
|
|
||||||
fi
|
|
||||||
if [ -z "${exitError}" ]; then
|
|
||||||
# exit cleanly
|
|
||||||
echo -e "${note}[$(stamp)] -- ${scriptName} completed" \
|
|
||||||
"--${normal}" >> "$logFile"
|
|
||||||
exit 0
|
|
||||||
else
|
|
||||||
# list generated errors and explanations then exit script with code 2
|
|
||||||
echo -e "\n${err}${scriptName} generated the following errors:" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
for errCode in "${exitError[@]}"; do
|
|
||||||
errStamp="${errCode%%_*}"
|
|
||||||
errValue="${errCode##*_}"
|
|
||||||
echo -e "${err}${errStamp} -- [ERROR] ${errorExplain[$errValue]}" \
|
|
||||||
"(code: ${errValue}) --${normal}" >> "$logFile"
|
|
||||||
done
|
|
||||||
exit 2
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
function checkExist {
|
|
||||||
if [ "$1" = "ff" ]; then
|
|
||||||
# find file
|
|
||||||
if [ -f "$2" ]; then
|
|
||||||
# found
|
|
||||||
return 0
|
|
||||||
else
|
|
||||||
# not found
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
elif [ "$1" = "fs" ]; then
|
|
||||||
# find file > 0 bytes
|
|
||||||
if [ -s "$2" ]; then
|
|
||||||
# found
|
|
||||||
return 0
|
|
||||||
else
|
|
||||||
# not found
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
elif [ "$1" = "fd" ]; then
|
|
||||||
# find directory
|
|
||||||
if [ -d "$2" ]; then
|
|
||||||
# found
|
|
||||||
return 0
|
|
||||||
else
|
|
||||||
# not found
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
### cleanup - cleanup files and directories created by this script
|
|
||||||
function cleanup {
|
|
||||||
## remove SQL dump file and directory
|
|
||||||
rm -rf "$sqlDumpDir" >> "$logFile" 2>&1
|
|
||||||
# verify directory is gone
|
|
||||||
checkExist fd "$sqlDumpDir"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
# directory still exists
|
|
||||||
exitWarn+=("[$(stamp)]_111")
|
|
||||||
else
|
|
||||||
# directory removed
|
|
||||||
echo -e "${op}[$(stamp)] Removed SQL temp directory${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
fi
|
|
||||||
|
|
||||||
## remove 503 error page
|
|
||||||
# check value of 'clean503' to see if this is necessary (=1) otherwise, skip
|
|
||||||
if [ "$clean503" = "1" ]; then
|
|
||||||
# proceed with cleanup
|
|
||||||
echo -e "${op}[$(stamp)] Removing 503 error page..." >> "$logFile"
|
|
||||||
rm -f "$webroot/$err503File" >> "$logFile" 2>&1
|
|
||||||
# verify file is actually gone
|
|
||||||
checkExist ff "$webroot/$err503File"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
# file still exists
|
|
||||||
exitWarn+=("[$(stamp)]_5030")
|
|
||||||
else
|
|
||||||
# file removed
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] 503 page removed from webroot" \
|
|
||||||
"--${normal}" >> "$logFile"
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
echo -e "${op}[$(stamp)] 503 error page never copied to webroot," \
|
|
||||||
"nothing to cleanup" >> "$logFile"
|
|
||||||
fi
|
|
||||||
|
|
||||||
## restart mailflow docker containers
|
|
||||||
# start and verify postfix
|
|
||||||
operateDocker start postfix
|
|
||||||
if [ "$dockerResultState" = "true" ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] Postfix container is running --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_103")
|
|
||||||
fi
|
|
||||||
# start and verify dovecot
|
|
||||||
operateDocker start dovecot
|
|
||||||
if [ "$dockerResultState" = "true" ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] Dovecot container is running --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_104")
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
### operate docker containers
|
|
||||||
function operateDocker {
|
|
||||||
containerName="$(docker ps -a --format '{{ .Names }}' --filter name=${COMPOSE_PROJECT_NAME}_${2}-mailcow_1)"
|
|
||||||
|
|
||||||
# determine action to take
|
|
||||||
if [ "$1" = "stop" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Stopping ${2}-mailcow container...${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
docker-compose stop --timeout ${dockerStopTimeout} ${2}-mailcow \
|
|
||||||
2>> "$logFile"
|
|
||||||
# verify container stopped (should return true)
|
|
||||||
dockerResultState="$(docker inspect -f '{{ .State.Running }}' $containerName)"
|
|
||||||
# verify clean stop (exit code 0)
|
|
||||||
dockerResultExit="$(docker inspect -f '{{ .State.ExitCode }}' $containerName)"
|
|
||||||
elif [ "$1" = "start" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Starting ${2}-mailcow container...${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
docker-compose start ${2}-mailcow 2>> "$logFile"
|
|
||||||
# verify
|
|
||||||
dockerResultState="$(docker inspect -f '{{ .State.Running }}' $containerName)"
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
### End of Functions ###
|
|
||||||
|
|
||||||
|
|
||||||
### Default parameters
|
|
||||||
|
|
||||||
# store the logfile in the same directory as this script using the script's name
|
|
||||||
# with the extension .log
|
|
||||||
scriptPath="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
|
|
||||||
scriptName="$( basename ${0} )"
|
|
||||||
logFile="$scriptPath/${scriptName%.*}.log"
|
|
||||||
|
|
||||||
# Set default docker-compose filename
|
|
||||||
dockerComposeFile=docker-compose.yml
|
|
||||||
|
|
||||||
# set default 503 error page name and location in scriptPath
|
|
||||||
err503Path="$scriptPath/503.html"
|
|
||||||
err503File="${err503Path##*/}"
|
|
||||||
|
|
||||||
# Set default docker-compose stop timeout
|
|
||||||
dockerStopTimeout=120
|
|
||||||
|
|
||||||
# Set default docker-compose start timeout
|
|
||||||
dockerStartTimeout=180
|
|
||||||
|
|
||||||
# set default borgDetails path to scriptPath
|
|
||||||
borgDetails="$scriptPath/mc_borg.details"
|
|
||||||
|
|
||||||
# set borg parameters to 'normal' verbosity
|
|
||||||
borgCreateParams='--stats'
|
|
||||||
borgPruneParams='--list'
|
|
||||||
|
|
||||||
|
|
||||||
### Set script parameters to null and initialize array variables
|
|
||||||
unset mailcowConfigFilePath
|
|
||||||
unset mailcowPath
|
|
||||||
unset dockerComposeFilePath
|
|
||||||
unset PARAMS
|
|
||||||
unset sqlDumpDir
|
|
||||||
unset webroot
|
|
||||||
unset clean503
|
|
||||||
unset borgXtra
|
|
||||||
unset borgExclude
|
|
||||||
unset borgPrune
|
|
||||||
unset BORG_BASE_DIR
|
|
||||||
unset BORG_RSH
|
|
||||||
unset BORG_REPO
|
|
||||||
unset BORG_PASSPHRASE
|
|
||||||
unset BORG_REMOTE_PATH
|
|
||||||
unset TMPDIR
|
|
||||||
unset dockerVolumeMail
|
|
||||||
unset dockerVolumeRspamd
|
|
||||||
unset dockerVolumePostfix
|
|
||||||
unset dockerVolumeRedis
|
|
||||||
unset dockerVolumeCrypt
|
|
||||||
exitError=()
|
|
||||||
errorExplain=()
|
|
||||||
exitWarn=()
|
|
||||||
warningExplain=()
|
|
||||||
borgConfig=()
|
|
||||||
xtraFiles=()
|
|
||||||
|
|
||||||
### Error codes
|
|
||||||
errorExplain[101]="Could not stop Postfix container. Please check docker logs"
|
|
||||||
errorExplain[102]="Could not stop Dovecot container. Please check docker logs"
|
|
||||||
errorExplain[103]="Could not start Postfix container. Please check docker logs"
|
|
||||||
errorExplain[104]="Could not start Dovecot container. Please check docker logs"
|
|
||||||
errorExplain[201]="There was a problem dumping the SQL database. It has NOT been backed up"
|
|
||||||
errorExplain[202]="There was a problem saving redis state information. It has NOT been backed up"
|
|
||||||
errorExplain[210]="Invalid or non-existant borg base directory specified (borg backup details file)"
|
|
||||||
errorExplain[211]="Invalid or non-existant path to borg SSH keyfile (borg backup details file)"
|
|
||||||
errorExplain[212]="Name of borg repo was not specified (borg backup details file)"
|
|
||||||
errorExplain[215]="Could not find/create 'tmp' directory within borg base directory. Please manually create it and ensure it's writable"
|
|
||||||
errorExplain[220]="Borg exited with a critical error. Please check this script's logfile for details"
|
|
||||||
errorExplain[221]="Borg prune exited with ERRORS. Please check this script's logfile for details"
|
|
||||||
|
|
||||||
|
|
||||||
### Warning codes & messages
|
|
||||||
warningExplain[111]="Could not remove SQL dump file and directory, please remove manually"
|
|
||||||
warningExplain[5030]="Could not remove 503 error page. This MUST be removed manually before NGINX will serve webclients!"
|
|
||||||
warningExplain[5031]="No webroot path was specified (-w parameter missing)"
|
|
||||||
warningExplain[5032]="The specified webroot (-w parameter) could not be found"
|
|
||||||
warningExplain[5033]="No 503 error page could be found. If not using the default located in the script directory, then check your -5 parameter"
|
|
||||||
warningExplain[5035]="Error copying 503 error page to webroot"
|
|
||||||
warn503="Web users will NOT be informed the server is down!"
|
|
||||||
warningExplain[2111]="No password used for SSH keys or access to remote borg repo. This is an insecure configuration"
|
|
||||||
warningExplain[2112]="No remote borg instance specified. Operations will be slower in this configuration"
|
|
||||||
warningExplain[2113]="The specified file containing extra files for inclusion in borgbackup could not be found"
|
|
||||||
warningExplain[2114]="The specified file containing exclusion patterns for borgbackup could not be found. Backup was performed as though NO exclusions were defined"
|
|
||||||
warningExplain[2115]="No paramters provided for borg prune. No repo pruning has taken place. You should reconsider this decision to control the size/history of your backups"
|
|
||||||
warningExplain[2116]="No additional locations are specified for inclusion in backup. ONLY Mailcow data and config files will be backed up (NO system files, etc). If this is unintentional, check the inclusion file referenced in your borgbackup settings"
|
|
||||||
warningExplain[2200]="Borg completed with warnings. Please check this script's logfile for details"
|
|
||||||
warningExplain[2201]="Borg exited with an unknown return-code. Please check this script's logfile for details"
|
|
||||||
warningExplain[2210]="Borg prune exited with warnings. Please check this script's logfile for details"
|
|
||||||
warningExplain[2212]="Borg prune exited with an unknown return-code. Please check this script's logfile for details"
|
|
||||||
|
|
||||||
|
|
||||||
### Process script parameters
|
|
||||||
|
|
||||||
# If parameters are provided but don't start with '-' then show the help page
|
|
||||||
# and exit with an error
|
|
||||||
if [ -n "$1" ] && [[ ! "$1" =~ ^- ]]; then
|
|
||||||
# show script help page
|
|
||||||
scriptHelp
|
|
||||||
fi
|
|
||||||
|
|
||||||
# use GetOpts to process parameters
|
|
||||||
while getopts ':l:v5:w:b:m:d:1:2:' PARAMS; do
|
|
||||||
case "$PARAMS" in
|
|
||||||
l)
|
|
||||||
# use provided location for logFile
|
|
||||||
logFile="${OPTARG%/}"
|
|
||||||
;;
|
|
||||||
v)
|
|
||||||
# verbose output from Borg
|
|
||||||
borgCreateParams='--list --stats'
|
|
||||||
borgPruneParams='--list'
|
|
||||||
;;
|
|
||||||
5)
|
|
||||||
# Full path to 503 error page
|
|
||||||
err503Path="${OPTARG%/}"
|
|
||||||
err503File="${err503Path##*/}"
|
|
||||||
;;
|
|
||||||
w)
|
|
||||||
# path to webserver webroot to copy 503 error page
|
|
||||||
webroot="${OPTARG%/}"
|
|
||||||
;;
|
|
||||||
b)
|
|
||||||
# path to file containing borgbackup settings and details
|
|
||||||
borgDetails="${OPTARG%/}"
|
|
||||||
;;
|
|
||||||
m)
|
|
||||||
# full path to mailcow.conf configuration file
|
|
||||||
mailcowConfigFilePath="${OPTARG%/}"
|
|
||||||
;;
|
|
||||||
d)
|
|
||||||
# name of docker-compose configuration file
|
|
||||||
dockerComposeFile="${OPTARG}"
|
|
||||||
;;
|
|
||||||
1)
|
|
||||||
# docker-compose stop timeout in seconds
|
|
||||||
dockerStopTimeout="${OPTARG}"
|
|
||||||
;;
|
|
||||||
2)
|
|
||||||
# docker-compose start timeout in seconds
|
|
||||||
dockerStartTimeout="${OPTARG}"
|
|
||||||
;;
|
|
||||||
?)
|
|
||||||
# unrecognized parameters trigger scriptHelp
|
|
||||||
scriptHelp
|
|
||||||
;;
|
|
||||||
esac
|
|
||||||
done
|
|
||||||
|
|
||||||
|
|
||||||
### Verify script pre-requisties
|
|
||||||
|
|
||||||
## If not running as root, display error on console and exit
|
|
||||||
if [ $(id -u) -ne 0 ]; then
|
|
||||||
echo -e "\n${err}This script MUST be run as ROOT. Exiting.${normal}"
|
|
||||||
exit 3
|
|
||||||
fi
|
|
||||||
|
|
||||||
## verify mailcow.conf location provided
|
|
||||||
if [ -z "$mailcowConfigFilePath" ]; then
|
|
||||||
echo -e "\n${err}You MUST provide the full path to your mailcow.conf" \
|
|
||||||
"configuration file. Exiting.${normal}"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
## verify mailcow.conf and extract path
|
|
||||||
if checkExist ff "$mailcowConfigFilePath"; then
|
|
||||||
# extract directory name
|
|
||||||
case $mailcowConfigFilePath in
|
|
||||||
*/*)
|
|
||||||
mailcowPath=${mailcowConfigFilePath%/*}
|
|
||||||
;;
|
|
||||||
*)
|
|
||||||
mailcowPath="."
|
|
||||||
;;
|
|
||||||
esac
|
|
||||||
else
|
|
||||||
echo -e "\n${err}Could not locate the specified mailcow configuration" \
|
|
||||||
"file: ${lit}${mailcowConfigFilePath}${normal}"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# verify docker-compose.yml exists at location of mailcow.conf (standard setup)
|
|
||||||
dockerComposeFilePath="${mailcowPath}/${dockerComposeFile}"
|
|
||||||
if ! checkExist ff "$dockerComposeFilePath"; then
|
|
||||||
echo -e "\n${err}Could not locate docker-compose configuration file:" \
|
|
||||||
"${lit}${dockerComposeFilePath}${normal}"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
## Ensure borgDetails file exists
|
|
||||||
checkExist ff "$borgDetails"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "1" ]; then
|
|
||||||
# sqlDetails file cannot be found
|
|
||||||
echo -e "\n${err}The file containing your borgbackup details does not" \
|
|
||||||
"exist (-b parameter)${normal}\n"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
### Log start of script operations
|
|
||||||
echo -e "${note}[$(stamp)] --- Start $scriptName execution ---${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] using ${lit}${mailcowConfigFilePath}" \
|
|
||||||
>> "$logFile"
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] using ${lit}${dockerComposeFilePath}" \
|
|
||||||
>> "$logFile"
|
|
||||||
|
|
||||||
|
|
||||||
### Import additional variables from mailcow configuration file
|
|
||||||
source "$mailcowConfigFilePath"
|
|
||||||
|
|
||||||
### Export PATH so this script can access all docker and docker-compose commands
|
|
||||||
export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
|
|
||||||
|
|
||||||
### Export logFile variable for use by Borg
|
|
||||||
export logFile="$logFile"
|
|
||||||
|
|
||||||
### Export docker container startup timeout variable
|
|
||||||
export COMPOSE_HTTP_TIMEOUT=${dockerStartTimeout}
|
|
||||||
|
|
||||||
## Get docker volume paths on filesystem
|
|
||||||
dockerVolumeMail=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_vmail-vol-1)
|
|
||||||
dockerVolumeRspamd=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_rspamd-vol-1)
|
|
||||||
dockerVolumePostfix=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_postfix-vol-1)
|
|
||||||
dockerVolumeRedis=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_redis-vol-1)
|
|
||||||
dockerVolumeCrypt=$(docker volume inspect -f '{{ .Mountpoint }}' ${COMPOSE_PROJECT_NAME}_crypt-vol-1)
|
|
||||||
|
|
||||||
|
|
||||||
### Create sqlDump temporary directory and sqlDumpFile name
|
|
||||||
sqlDumpDir=$( mktemp -d )
|
|
||||||
sqlDumpFile="backup-`date +%Y%m%d_%H%M%S`.sql"
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] mySQL dump file will be stored" \
|
|
||||||
"at: ${lit}${sqlDumpDir}/${sqlDumpFile}${normal}" >> "$logFile"
|
|
||||||
|
|
||||||
|
|
||||||
### 503 error page: If you dont' plan on using the auto-copied 503 then comment
|
|
||||||
### this entire section starting with '--- Begin 503 section ---' until
|
|
||||||
### '--- End 503 section ---' to suppress generated warnings
|
|
||||||
|
|
||||||
### --- Begin 503 section ---
|
|
||||||
|
|
||||||
## Check if webroot has been specified, if not, skip this entire section since there is nowhere to copy the 503 file.
|
|
||||||
if [ -z "$webroot" ]; then
|
|
||||||
# no webroot path provided
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] ${warn503} --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
exitWarn+=("[$(stamp)]_5031")
|
|
||||||
clean503=0
|
|
||||||
else
|
|
||||||
# verify webroot actually exists
|
|
||||||
checkExist fd "$webroot"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "1" ]; then
|
|
||||||
# webroot directory specified could not be found
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] ${warn503} --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
exitWarn+=("[$(stamp)]_5032")
|
|
||||||
clean503=0
|
|
||||||
else
|
|
||||||
# webroot exists
|
|
||||||
echo -e "${op}[$(stamp)] Using webroot: ${lit}${webroot}${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
# Verify 503 file existance at given path
|
|
||||||
checkExist ff "$err503Path"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "1" ]; then
|
|
||||||
# 503 file could not be found
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] ${warn503} --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
exitWarn+=("[$(stamp)]_5033")
|
|
||||||
clean503=0
|
|
||||||
else
|
|
||||||
# 503 file exists and webroot is valid. Let's copy it!
|
|
||||||
echo -e "${op}[$(stamp)] ${err503File} found at ${lit}${err503Path}" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
echo -e "${op}[$(stamp)] Copying 503 error page to webroot..." \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
cp "${err503Path}" "$webroot/" >> "$logFile" 2>&1
|
|
||||||
copyResult="$?"
|
|
||||||
# verify copy was successful
|
|
||||||
if [ "$copyResult" = "1" ]; then
|
|
||||||
# copy was unsuccessful
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] ${warn503} --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
exitWarn+=("[$(stamp)]_5035")
|
|
||||||
clean503=0
|
|
||||||
else
|
|
||||||
# copy was successful
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] 503 error page" \
|
|
||||||
"successfully copied to webroot --${normal}" >> "$logFile"
|
|
||||||
clean503=1
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
### --- End 503 section ---
|
|
||||||
|
|
||||||
|
|
||||||
### Change directory to mailcowPath
|
|
||||||
cd "$mailcowPath"
|
|
||||||
|
|
||||||
|
|
||||||
### Stop postfix and dovecot so mailflow is stopped until backup is completed
|
|
||||||
## Stop postfix-mailcow container
|
|
||||||
operateDocker stop postfix
|
|
||||||
# process result
|
|
||||||
if [ "$dockerResultState" = "false" ] && [ "$dockerResultExit" -eq 0 ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] Postfix container stopped --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_101")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
fi
|
|
||||||
## Stop dovecot-mailcow container
|
|
||||||
operateDocker stop dovecot
|
|
||||||
# process result
|
|
||||||
if [ "$dockerResultState" = "false" ] && [ "$dockerResultExit" -eq 0 ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] Dovecot container stopped --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_102")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
### Dump SQL
|
|
||||||
echo -e "${op}[$(stamp)] Dumping mailcow SQL database...${normal}" >> "$logFile"
|
|
||||||
docker-compose exec -T mysql-mailcow mysqldump --default-character-set=utf8mb4 \
|
|
||||||
-u${DBUSER} -p${DBPASS} ${DBNAME} > "$sqlDumpDir/$sqlDumpFile" \
|
|
||||||
2>> "$logFile"
|
|
||||||
dumpResult=$(docker-compose exec -T mysql-mailcow echo "$?")
|
|
||||||
## very mysqldump completed successfully
|
|
||||||
if [ "$dumpResult" = "0" ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] mySQLdump completed successfully --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_201")
|
|
||||||
fi
|
|
||||||
## verify the dump file was actually written to disk
|
|
||||||
checkExist fs "$sqlDumpDir/$sqlDumpFile"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
echo -e "${ok}[$(stamp)] -- [SUCCESS] SQL successfully dumped --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_201")
|
|
||||||
fi
|
|
||||||
|
|
||||||
### Save redis state
|
|
||||||
## Delete any existing redis dump file otherwise our file check will be useless
|
|
||||||
echo -e "${op}[$(stamp)] Cleaning up old redis state backup...${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
checkExist ff "$dockerVolumeRedis/dump.rdb"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
echo -e "${lit}[$(stamp)] Old redis backup found. ${op}Deleting..." \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
rm -f "$dockerVolumeRedis/dump.rdb" 2>> "$logFile"
|
|
||||||
echo -e "${op}[$(stamp)] ...done${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
echo -e "${op}[$(stamp)] No old redis backup found${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
fi
|
|
||||||
## Export redis
|
|
||||||
echo -e "${op}[$(stamp)] Saving redis state information...${normal}" >> "$logFile"
|
|
||||||
docker-compose exec -T redis-mailcow redis-cli save >> "$logFile" 2>&1
|
|
||||||
saveResult=$(docker-compose exec -T redis-mailcow echo "$?")
|
|
||||||
# verify save operation completed successfully
|
|
||||||
if [ "$saveResult" = "0" ]; then
|
|
||||||
echo -e "${info}[$(stamp)] -- [INFO] redis save-state successful --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_202")
|
|
||||||
fi
|
|
||||||
## verify save-file written to disk
|
|
||||||
checkExist fs "$dockerVolumeRedis/dump.rdb"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
echo -e "${ok}[$(stamp)] -- [SUCCESS] redis state saved --${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
else
|
|
||||||
exitError+=("[$(stamp)]_202")
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
### Call borgbackup to copy actual files
|
|
||||||
echo -e "${op}[$(stamp)] Pre-backup tasks completed, calling borgbackup..." \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
|
|
||||||
## Get borgbackup settings and repo details
|
|
||||||
# read definition file and map to array variable
|
|
||||||
mapfile -t borgConfig < "$borgDetails"
|
|
||||||
## check if any required borg configuration variables in defintion file are
|
|
||||||
## empty and exit with error, otherwise, map array items to variables
|
|
||||||
# check: borg base directory
|
|
||||||
echo -e "${op}[$(stamp)] Verifying supplied borg configuration variables..." \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
if [ -z "${borgConfig[0]}" ]; then
|
|
||||||
exitError+=("[$(stamp)]_210")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
else
|
|
||||||
# verify the path actually exists
|
|
||||||
checkExist fd "${borgConfig[0]}"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "1" ]; then
|
|
||||||
# borg base directory specified could not be found
|
|
||||||
exitError+=("[$(stamp)]_210")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
fi
|
|
||||||
echo -e "${op}[$(stamp)] Borg base dir... OK${normal}" >> "$logFile"
|
|
||||||
export BORG_BASE_DIR="${borgConfig[0]%/}"
|
|
||||||
fi
|
|
||||||
# check: path to SSH keyfile
|
|
||||||
if [ -z "${borgConfig[1]}" ]; then
|
|
||||||
exitError+=("[$(stamp)]_211")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
else
|
|
||||||
checkExist ff "${borgConfig[1]}"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = 1 ]; then
|
|
||||||
# SSH keyfile specified could not be found
|
|
||||||
exitError+=("[$(stamp)]_211")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
fi
|
|
||||||
echo -e "${op}[$(stamp)] Borg SSH key... OK${normal}" >> "$logFile"
|
|
||||||
export BORG_RSH="ssh -i ${borgConfig[1]}"
|
|
||||||
fi
|
|
||||||
# check: name of borg repo
|
|
||||||
if [ -z "${borgConfig[2]}" ]; then
|
|
||||||
exitError+=("[$(stamp)]_212")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
else
|
|
||||||
echo -e "${op}[$(stamp)] Borg REPO name... OK${normal}" >> "$logFile"
|
|
||||||
export BORG_REPO="${borgConfig[2]}"
|
|
||||||
fi
|
|
||||||
# repo password
|
|
||||||
if [ -n "${borgConfig[3]}" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Borg SSH/REPO password... OK${normal}" >> "$logFile"
|
|
||||||
export BORG_PASSPHRASE="${borgConfig[3]}"
|
|
||||||
else
|
|
||||||
exitWarn+=("[$(stamp)]_2111")
|
|
||||||
# if the password was omitted by mistake, export a dummy password so borg
|
|
||||||
# fails with an error instead of sitting and waiting for input
|
|
||||||
export BORG_PASSPHRASE="DummyPasswordSoBorgFails"
|
|
||||||
fi
|
|
||||||
# additional files to be backed up
|
|
||||||
borgXtra="${borgConfig[4]}"
|
|
||||||
# file with pattern definition for excluded files
|
|
||||||
borgExclude="${borgConfig[5]}"
|
|
||||||
# parameters for borg prune
|
|
||||||
borgPrune="${borgConfig[6]}"
|
|
||||||
# export: borg remote path (if not blank)
|
|
||||||
if [ -n "${borgConfig[7]}" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Borg REMOTE path... OK${normal}" >> "$logFile"
|
|
||||||
export BORG_REMOTE_PATH="${borgConfig[7]}"
|
|
||||||
else
|
|
||||||
exitWarn+=("[$(stamp)]_2112")
|
|
||||||
fi
|
|
||||||
|
|
||||||
## If borgXtra exists, map contents to an array variable
|
|
||||||
if [ -n "$borgXtra" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Processing referenced extra files list for" \
|
|
||||||
"borgbackup to include in backup${normal}" >> "$logFile"
|
|
||||||
checkExist ff "$borgXtra"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Found ${lit}${borgXtra}${normal}" >> "$logFile"
|
|
||||||
mapfile -t xtraFiles < "$borgXtra"
|
|
||||||
echo -e "${op}[$(stamp)] Processed extra files list for inclusion in" \
|
|
||||||
"borgbackup${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
exitWarn+=("[$(stamp)]_2113")
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
# no extra locations specified
|
|
||||||
echo -e "${op}[$(stamp)] No additional locations specified for backup." \
|
|
||||||
"Only Mailcow data and config files will be backed up.${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
exitWarn+=("[$(stamp)]_2116")
|
|
||||||
fi
|
|
||||||
|
|
||||||
## Check if borgExclude exists since borg will throw an error if it's missing
|
|
||||||
if [ -n "$borgExclude" ]; then
|
|
||||||
checkExist ff "$borgExclude"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
echo -e "${op}[$(stamp)] Found ${lit}${borgExclude}${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
else
|
|
||||||
# file not found, unset the variable so it's like it was not specified
|
|
||||||
# in the first place and continue with backup
|
|
||||||
unset borgExclude
|
|
||||||
exitWarn+=("[$(stamp)]_2114")
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
echo -e "${op}[$(stamp)] Exclusion pattern file not specified." \
|
|
||||||
"No exclusions will be processed${normal}" >> "$logFile"
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
## Export TMPDIR environment variable for borg via python
|
|
||||||
## Python requires a writable temporary directory when unpacking borg and
|
|
||||||
## executing commands. This defaults to /tmp but many systems mount /tmp with
|
|
||||||
## the 'noexec' option for security. Thus, we will use/create a 'tmp' folder
|
|
||||||
## within the BORG_BASE_DIR and instruct python to use that instead of /tmp
|
|
||||||
# check if BORG_BASE_DIR/tmp exists, if not, create it
|
|
||||||
echo -e "${op}[$(stamp)] Checking for tmp directory at ${lit}${BORG_BASE_DIR}" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
checkExist fd "$BORG_BASE_DIR/tmp"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "1" ]; then
|
|
||||||
# folder not found
|
|
||||||
echo -e "${op}[$(stamp)] tmp folder not found... creating${lit}" \
|
|
||||||
"${BORG_BASE_DIR}/tmp${normal}" >> "$logFile"
|
|
||||||
mkdir "$BORG_BASE_DIR/tmp" 2>> "$logFile"
|
|
||||||
# verify folder created
|
|
||||||
checkExist fd "$BORG_BASE_DIR/tmp"
|
|
||||||
checkResult="$?"
|
|
||||||
if [ "$checkResult" = "0" ]; then
|
|
||||||
# folder exists
|
|
||||||
echo -e "${op}[$(stamp)] tmp folder created within borg base directory" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
else
|
|
||||||
# problem creating folder and script will exit
|
|
||||||
exitError+=("[$(stamp)]_215")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
# folder found
|
|
||||||
echo -e "${op}[$(stamp)] tmp folder found within borg base directory" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
fi
|
|
||||||
# export TMPDIR environment variable
|
|
||||||
export TMPDIR="${BORG_BASE_DIR}/tmp"
|
|
||||||
|
|
||||||
|
|
||||||
## Generate and execute borg
|
|
||||||
# commandline depends on whether borgExclude is empty or not
|
|
||||||
if [ -z "$borgExclude" ]; then
|
|
||||||
# borgExclude is empty
|
|
||||||
echo -e "${bold}${op}[$(stamp)] Executing borg without exclusions${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
borg --show-rc create ${borgCreateParams} ::`date +%Y-%m-%d_%H%M%S` \
|
|
||||||
"${xtraFiles[@]}" \
|
|
||||||
"${sqlDumpDir}" \
|
|
||||||
"${dockerVolumeMail}" "${dockerVolumeRspamd}" "${dockerVolumePostfix}" \
|
|
||||||
"${dockerVolumeRedis}" "${dockerVolumeCrypt}" \
|
|
||||||
2>> "$logFile"
|
|
||||||
else
|
|
||||||
# borgExclude is not empty
|
|
||||||
echo -e "${bold}${op}[$(stamp)] Executing borg with exclusions${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
borg --show-rc create ${borgCreateParams} --exclude-from "${borgExclude}" \
|
|
||||||
::`date +%Y-%m-%d_%H%M%S` \
|
|
||||||
"${xtraFiles[@]}" \
|
|
||||||
"${sqlDumpDir}" \
|
|
||||||
"${dockerVolumeMail}" "${dockerVolumeRspamd}" "${dockerVolumePostfix}" \
|
|
||||||
"${dockerVolumeRedis}" "${dockerVolumeCrypt}" \
|
|
||||||
2>> "$logFile"
|
|
||||||
fi
|
|
||||||
|
|
||||||
## Check status of borg operation
|
|
||||||
borgResult="$?"
|
|
||||||
if [ "$borgResult" -eq 0 ]; then
|
|
||||||
echo -e "${ok}[$(stamp)] -- [SUCCESS] Borg backup completed successfully --" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
elif [ "$borgResult" -eq 1 ]; then
|
|
||||||
exitWarn+=("[$(stamp)]_2200")
|
|
||||||
elif [ "$borgResult" -ge 2 ]; then
|
|
||||||
exitError+=("[$(stamp)]_220")
|
|
||||||
cleanup
|
|
||||||
quit
|
|
||||||
else
|
|
||||||
exitWarn+=("[$(stamp)]_2201")
|
|
||||||
fi
|
|
||||||
|
|
||||||
## Generate and execute borg prune
|
|
||||||
# command depends on whether or not parameters have been defined
|
|
||||||
if [ -n "$borgPrune" ]; then
|
|
||||||
# parameters defined
|
|
||||||
echo -e "${bold}${op}[$(stamp)] Executing borg prune operation${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
borg prune --show-rc -v ${borgPruneParams} ${borgPrune} \
|
|
||||||
2>> "$logFile"
|
|
||||||
# check return-status
|
|
||||||
pruneResult="$?"
|
|
||||||
if [ "$pruneResult" -eq 0 ]; then
|
|
||||||
echo -e "${ok}[$(stamp)] -- [SUCCESS] Borg prune completed successfully" \
|
|
||||||
"--${normal}" >> "$logFile"
|
|
||||||
elif [ "$pruneResult" -eq 1 ]; then
|
|
||||||
exitWarn+=("[$(stamp)]_2210")
|
|
||||||
elif [ "$pruneResult" -ge 2 ]; then
|
|
||||||
exitError+=("[$(stamp)]_221")
|
|
||||||
else
|
|
||||||
exitWarn+=("[$(stamp)]_2212")
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
# parameters not defined... skip pruning
|
|
||||||
exitWarn+=("[$(stamp)]_2115")
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
### borgbackup completed
|
|
||||||
echo -e "${op}[$(stamp)] Borgbackup completed... begin cleanup" \
|
|
||||||
"${normal}" >> "$logFile"
|
|
||||||
|
|
||||||
|
|
||||||
### Exit script
|
|
||||||
echo -e "${bold}${op}[$(stamp)] ***Normal exit process***${normal}" \
|
|
||||||
>> "$logFile"
|
|
||||||
cleanup
|
|
||||||
echo -e "${bold}${ok}[$(stamp)] -- [SUCCESS] All processes completed" \
|
|
||||||
"successfully --${normal}" >> "$logFile"
|
|
||||||
quit
|
|
||||||
|
|
||||||
# This code should not be executed since the 'quit' function should terminate
|
|
||||||
# this script. Therefore, exit with code 99 if we get to this point.
|
|
||||||
exit 99
|
|
@ -1,8 +0,0 @@
|
|||||||
<path to borgbackup base directory> /var/borgbackup
|
|
||||||
<path to SSH private key for remote server> /var/borgbackup/sshPrivate.key
|
|
||||||
<connection string to remote repo> user@servername.tld:repoName/
|
|
||||||
<password for repo> pAsSwOrd
|
|
||||||
<path to file listing extra files> /root/scripts/xtraLocations.borg
|
|
||||||
<path to file with exclusions> /root/scripts/excludeLocations.borg
|
|
||||||
<purge timeframe options> --keep-within=7d --keep-daily=30 --keep-weekly=12 --keep-monthly=-1
|
|
||||||
<location of borg remote instance> borg1
|
|
@ -1,27 +0,0 @@
|
|||||||
/etc/network/interfaces
|
|
||||||
/etc/network/interfaces.d/
|
|
||||||
/etc/systemd/timesyncd.conf
|
|
||||||
/etc/profile
|
|
||||||
/etc/bash.bashrc
|
|
||||||
/etc/skel/
|
|
||||||
/etc/nanorc
|
|
||||||
/etc/msmtprc
|
|
||||||
/etc/msmtp_aliases
|
|
||||||
/etc/apt/sources.list
|
|
||||||
/etc/apt/sources.list.d/
|
|
||||||
/etc/apt/listchanges.conf
|
|
||||||
/etc/apt/apt.conf.d/50unattended-upgrades
|
|
||||||
/etc/apt/apt.conf.d/20auto-upgrades
|
|
||||||
/etc/ssh/
|
|
||||||
/etc/logwatch/
|
|
||||||
/etc/letsencrypt/
|
|
||||||
/etc/fail2ban/fail2ban.local
|
|
||||||
/etc/fail2ban/jail.local
|
|
||||||
/etc/fail2ban/action.d/ufw.conf
|
|
||||||
/etc/fail2ban/filter.d/ufw-probe.conf
|
|
||||||
/etc/fail2ban/jail.d/
|
|
||||||
/root/.bashrc
|
|
||||||
/root/scripts/
|
|
||||||
/root/.ssh/
|
|
||||||
/etc/nginx/
|
|
||||||
/usr/share/nginx/html/
|
|
Loading…
Reference in New Issue
Block a user