Parcourir la source

Run update.sh

Docker Library Bot il y a 8 ans
Parent
commit
dda8c21160
11 fichiers modifiés avec 201 ajouts et 119 suppressions
  1. 29 19
      adminer/README.md
  2. 6 6
      alpine/README.md
  3. 2 2
      chronograf/README.md
  4. 29 13
      convertigo/README.md
  5. 6 2
      eclipse-mosquitto/README.md
  6. 4 3
      fsharp/README.md
  7. 4 2
      neo4j/README.md
  8. 6 5
      ros/README.md
  9. 73 53
      silverpeas/README.md
  10. 27 9
      storm/README.md
  11. 15 5
      zookeeper/README.md

+ 29 - 19
adminer/README.md

@@ -58,7 +58,9 @@ Adminer (formerly phpMinAdmin) is a full-featured database management tool writt
 
 ### Standalone
 
-	$ docker run --link some_database:db -p 8080:8080 adminer
+```console
+$ docker run --link some_database:db -p 8080:8080 adminer
+```
 
 Then you can hit `http://localhost:8080` or `http://host-ip:8080` in your browser.
 
@@ -66,7 +68,9 @@ Then you can hit `http://localhost:8080` or `http://host-ip:8080` in your browse
 
 If you are already running a FastCGI capable web server you might prefer running adminer via FastCGI:
 
-	$ docker run --link some_database:db -p 9000:9000 adminer:fastcgi
+```console
+$ docker run --link some_database:db -p 9000:9000 adminer:fastcgi
+```
 
 Then point your web server to port 9000 of the container.
 
@@ -78,25 +82,29 @@ This image bundles all official adminer plugins. You can find the list of plugin
 
 To load plugins you can pass a list of filenames in `ADMINER_PLUGINS`:
 
-	$ docker run --link some_database:db -p 8080:8080 -e ADMINER_PLUGINS='tables-filter tinymce' adminer
+```console
+$ docker run --link some_database:db -p 8080:8080 -e ADMINER_PLUGINS='tables-filter tinymce' adminer
+```
 
 If a plugin *requires* parameters to work correctly you will need to add a custom file to the container:
 
-	$ docker run --link some_database:db -p 8080:8080 -e ADMINER_PLUGINS='login-servers' adminer
-	Unable to load plugin file "login-servers", because it has required parameters: servers
-	Create a file "/var/www/html/plugins-enabled/login-servers.php" with the following contents to load the plugin:
-	
-	<?php
-	require_once('plugins/login-servers.php');
-	
-	/** Set supported servers
-	    * @param array array($domain) or array($domain => $description) or array($category => array())
-	    * @param string
-	    */
-	return new AdminerLoginServers(
-	    $servers = ???,
-	    $driver = 'server'
-	);
+```console
+$ docker run --link some_database:db -p 8080:8080 -e ADMINER_PLUGINS='login-servers' adminer
+Unable to load plugin file "login-servers", because it has required parameters: servers
+Create a file "/var/www/html/plugins-enabled/login-servers.php" with the following contents to load the plugin:
+
+<?php
+require_once('plugins/login-servers.php');
+
+/** Set supported servers
+    * @param array array($domain) or array($domain => $description) or array($category => array())
+    * @param string
+    */
+return new AdminerLoginServers(
+    $servers = ???,
+    $driver = 'server'
+);
+```
 
 To load a custom plugin you can add PHP scripts that return the instance of the plugin object to `/var/www/html/plugins-enabled/`.
 
@@ -106,7 +114,9 @@ The image bundles all the designs that are available in the source package of ad
 
 To use a bundled design you can pass its name in `ADMINER_DESIGN`:
 
-	$ docker run --link some_database:db -p 8080:8080 -e ADMINER_DESIGN='nette' adminer
+```console
+$ docker run --link some_database:db -p 8080:8080 -e ADMINER_DESIGN='nette' adminer
+```
 
 To use a custom design you can add a file called `/var/www/html/adminer.css`.
 

+ 6 - 6
alpine/README.md

@@ -16,12 +16,12 @@ WARNING:
 
 # Supported tags and respective `Dockerfile` links
 
--	[`3.1` (*versions/library-3.1/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/cb0a5a01230693785e876a91b3484c4c05b2b06b/versions/library-3.1/Dockerfile)
--	[`3.2` (*versions/library-3.2/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/c60ae198ab8ac02a28725dfe990b1742c583b156/versions/library-3.2/Dockerfile)
--	[`3.3` (*versions/library-3.3/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/401eef8812da24776468ff885a41afd2d7958882/versions/library-3.3/Dockerfile)
--	[`3.4` (*versions/library-3.4/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/47a904af6b2702a947fe3ce7d741f877c30c11d0/versions/library-3.4/Dockerfile)
--	[`3.5`, `latest` (*versions/library-3.5/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/c7368b846ee805b286d9034a39e0bbf40bc079b3/versions/library-3.5/Dockerfile)
--	[`edge` (*versions/library-edge/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/60d6d06885459a9fa7e5d51d986be89ce61d9c41/versions/library-edge/Dockerfile)
+-	[`3.1` (*versions/library-3.1/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/3d0e79188adb609440a3b2c8516d7bd0b9db9107/versions/library-3.1/Dockerfile)
+-	[`3.2` (*versions/library-3.2/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/ef2b09086c8b3f9f673c5100470462a28efe9038/versions/library-3.2/Dockerfile)
+-	[`3.3` (*versions/library-3.3/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/84063540fc7568984d0c1f819107fd02081696c7/versions/library-3.3/Dockerfile)
+-	[`3.4` (*versions/library-3.4/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/fb9930f2c3179e629f5cff887742f31bbb7d8d46/versions/library-3.4/Dockerfile)
+-	[`3.5`, `latest` (*versions/library-3.5/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/cb5e568b84f4a48bb1245c602bae3de38a67bd55/versions/library-3.5/Dockerfile)
+-	[`edge` (*versions/library-edge/Dockerfile*)](https://github.com/gliderlabs/docker-alpine/blob/fa6353d492aefa36c4a2e1c6130b891c85047e75/versions/library-edge/Dockerfile)
 
 # Quick reference
 

+ 2 - 2
chronograf/README.md

@@ -46,7 +46,7 @@ WARNING:
 
 # Chronograf
 
-Chronograf is a simple to install graphing and visualization application that you deploy behind your firewall to perform ad-hoc exploration of your InfluxDB data. It includes support for templates and a library of intelligent, pre-configured dashboards for common data sets.
+Chronograf is InfluxData’s open source web application. Use Chronograf with the other components of the [TICK](https://www.influxdata.com/products/) stack for infrastructure monitoring, alert management, data visualization, and database management.
 
 ![logo](https://raw.githubusercontent.com/docker-library/docs/43d87118415bb75d7bb107683e79cd6d69186f67/chronograf/logo.png)
 
@@ -108,7 +108,7 @@ Try combining this with Telegraf to get dashboards for your infrastructure withi
 
 ## Official Documentation
 
-See the [official docs](https://docs.influxdata.com/chronograf/latest/introduction/getting_started/) for information on creating visualizations.
+See the [official docs](https://docs.influxdata.com/chronograf/latest/) for information on creating visualizations.
 
 # Image Variants
 

+ 29 - 13
convertigo/README.md

@@ -63,7 +63,9 @@ Convertigo Community edition brought to you by Convertigo SA (Paris & San Franci
 
 ## Quick start
 
-	$ docker run --name C8O -d -p 28080:28080 convertigo
+```console
+$ docker run --name C8O -d -p 28080:28080 convertigo
+```
 
 This will start a container running the minimum Convertigo MBaaS server. Convertigo MBaaS uses images' **/workspace** directory to store configuration file and deployed projects as an Docker volume.
 
@@ -75,11 +77,15 @@ Convertigo MBaaS FullSync module uses Apache CouchDB 1.6.1 as NoSQL repository.
 
 Launch CouchDB container and name it 'fullsync'
 
-	docker run -d --name fullsync couchdb:1.6.1
+```console
+$ docker run -d --name fullsync couchdb:1.6.1
+```
 
 Then launch Convertigo and link it to the running 'fullsync' container. Convertigo MBaaS sever will automatically use it as its fullsync repository.
 
-	docker run -d --name C8O-MBAAS --link fullsync:couchdb -p 28080:28080 convertigo
+```console
+$ docker run -d --name C8O-MBAAS --link fullsync:couchdb -p 28080:28080 convertigo
+```
 
 ## Link Convertigo to a Billing & Analytics database
 
@@ -87,18 +93,22 @@ Then launch Convertigo and link it to the running 'fullsync' container. Converti
 
 MySQL is the recommended database for holding Convertigo MBaaS server analytics. You can use this command to run convertigo and link it to a running MySQL container. Change `[mysql-container]` to the container name, and `[username for the c8oAnalytics db]`, `[password for specified db user]` with the values for your MySQL configuration.
 
-	docker run -d --name C8O-MBAAS --link [mysql-container]:mysql -p 28080:28080                         \
-	    -e JAVA_OPTS="-Dconvertigo.engine.billing.enabled=true                                           \ 
-	            -Dconvertigo.engine.billing.persistence.jdbc.username=[username for the c8oAnalytics db] \
-	            -Dconvertigo.engine.billing.persistence.jdbc.password=[password for specified db user]   \
-	            -Dconvertigo.engine.billing.persistence.jdbc.url=jdbc:mysql://mysql:3306/c8oAnalytics"   \
-	convertigo
+```console
+$ docker run -d --name C8O-MBAAS --link [mysql-container]:mysql -p 28080:28080                         \
+    -e JAVA_OPTS="-Dconvertigo.engine.billing.enabled=true                                           \ 
+            -Dconvertigo.engine.billing.persistence.jdbc.username=[username for the c8oAnalytics db] \
+            -Dconvertigo.engine.billing.persistence.jdbc.password=[password for specified db user]   \
+            -Dconvertigo.engine.billing.persistence.jdbc.url=jdbc:mysql://mysql:3306/c8oAnalytics"   \
+convertigo
+```
 
 ## Where is Convertigo MBaaS server storing deployed projects
 
 Projects are deployed in the Convertigo workspace, a simple file system directory. You can map the docker container **/workspace** to your physical system by using :
 
-	docker run --name C8O-MBAAS -v $(pwd):/workspace -d -p 28080:28080 convertigo
+```console
+$ docker run --name C8O-MBAAS -v $(pwd):/workspace -d -p 28080:28080 convertigo
+```
 
 You can share the same workspace by all Convertigo containers. This this case, when you deploy a project on a Convertigo container, it will be seen by others. This is the best way to build multi-instance load balanced Convertigo server farms.
 
@@ -118,13 +128,17 @@ These accounts can be configured through the *administration console* and saved
 
 You can change the default administration account :
 
-	docker run -d --name C8O-MBAAS -e CONVERTIGO_ADMIN_USER=administrator -e CONVERTIGO_ADMIN_PASSWORD=s3cret -p 28080:28080 convertigo
+```console
+$ docker run -d --name C8O-MBAAS -e CONVERTIGO_ADMIN_USER=administrator -e CONVERTIGO_ADMIN_PASSWORD=s3cret -p 28080:28080 convertigo
+```
 
 ### `CONVERTIGO_TESTPLATFORM_USER` and `CONVERTIGO_TESTPLATFORM_PASSWORD` variables
 
 You can lock the **testplatform** by setting the account :
 
-	docker run -d --name C8O-MBAAS -e CONVERTIGO_TESTPLATFORM_USER=tp_user -e CONVERTIGO_TESTPLATFORM_PASSWORD=s3cret -p 28080:28080 convertigo
+```console
+$ docker run -d --name C8O-MBAAS -e CONVERTIGO_TESTPLATFORM_USER=tp_user -e CONVERTIGO_TESTPLATFORM_PASSWORD=s3cret -p 28080:28080 convertigo
+```
 
 ## `JAVA_OPTS` Environment variable
 
@@ -132,7 +146,9 @@ Convertigo is based on a *Java* process with some defaults *JVM* options. You ca
 
 Add any *Java JVM* options such as -Xmx or -D[something]
 
-	docker run -d --name C8O-MBAAS -e JAVA_OPTS="-Xmx4096 -DjvmRoute=server1" -p 28080:28080 convertigo
+```console
+$ docker run -d --name C8O-MBAAS -e JAVA_OPTS="-Xmx4096 -DjvmRoute=server1" -p 28080:28080 convertigo
+```
 
 ## Pre configurated Docker compose stack
 

+ 6 - 2
eclipse-mosquitto/README.md

@@ -65,7 +65,9 @@ Three directories have been created in the image to be used for configuration, p
 
 When running the image, the default configuration values are used. To use a custom configuration file, mount a **local** configuration file to `/mosquitto/config/mosquitto.conf`
 
-	docker run -it -p 1883:1883 -p 9001:9001 -v mosquitto.conf:/mosquitto/config/mosquitto.conf eclipse-mosquitto
+```console
+$ docker run -it -p 1883:1883 -p 9001:9001 -v mosquitto.conf:/mosquitto/config/mosquitto.conf eclipse-mosquitto
+```
 
 Configuration can be changed to:
 
@@ -84,7 +86,9 @@ i.e. add the following to `mosquitto.conf`:
 
 Run a container using the new image:
 
-	docker run -it -p 1883:1883 -p 9001:9001 -v mosquitto.conf:/mosquitto/config/mosquitto.conf -v /mosquitto/data -v /mosquitto/log eclipse-mosquitto
+```console
+$ docker run -it -p 1883:1883 -p 9001:9001 -v mosquitto.conf:/mosquitto/config/mosquitto.conf -v /mosquitto/data -v /mosquitto/log eclipse-mosquitto
+```
 
 **Note**: if the mosquitto configuration (mosquitto.conf) was modified to use non-default ports, the docker run command will need to be updated to expose the ports that have been configured.
 

+ 4 - 3
fsharp/README.md

@@ -16,9 +16,10 @@ WARNING:
 
 # Supported tags and respective `Dockerfile` links
 
--	[`latest`, `4`, `4.1`, `4.1.0.1` (*4.1.0.1/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/289a5e0066c69255c055dd5ee22f2e583d37fab4/4.1.0.1/Dockerfile)
--	[`4.0`, `4.0.1`, `4.0.1.1` (*4.0.1.1/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/289a5e0066c69255c055dd5ee22f2e583d37fab4/4.0.1.1/Dockerfile)
--	[`4.0.0.4` (*4.0.0.4/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/289a5e0066c69255c055dd5ee22f2e583d37fab4/4.0.0.4/Dockerfile)
+-	[`latest`, `4`, `4.1`, `4.1.18` (*4.1.18/mono/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/a28196740e38035beea04c41d5862136413281e3/4.1.18/mono/Dockerfile)
+-	[`4.1.0.1` (*4.1.0.1/mono/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/a28196740e38035beea04c41d5862136413281e3/4.1.0.1/mono/Dockerfile)
+-	[`4.0`, `4.0.1`, `4.0.1.1` (*4.0.1.1/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/a28196740e38035beea04c41d5862136413281e3/4.0.1.1/Dockerfile)
+-	[`4.0.0.4` (*4.0.0.4/Dockerfile*)](https://github.com/fsprojects/docker-fsharp/blob/a28196740e38035beea04c41d5862136413281e3/4.0.0.4/Dockerfile)
 
 # Quick reference
 

+ 4 - 2
neo4j/README.md

@@ -16,8 +16,10 @@ WARNING:
 
 # Supported tags and respective `Dockerfile` links
 
--	[`3.1.4`, `3.1`, `latest` (*3.1.4/community/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/86c2eb04e9f44e2c6febf5c14d0a7fe67c81f35c/3.1.4/community/Dockerfile)
--	[`3.1.4-enterprise`, `3.1-enterprise`, `enterprise` (*3.1.4/enterprise/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/86c2eb04e9f44e2c6febf5c14d0a7fe67c81f35c/3.1.4/enterprise/Dockerfile)
+-	[`3.2.0`, `3.2`, `latest` (*3.2.0/community/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/c0364e1bcbc35a536abb93ec88dc3bfc288b23bf/3.2.0/community/Dockerfile)
+-	[`3.2.0-enterprise`, `3.2-enterprise`, `enterprise` (*3.2.0/enterprise/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/c0364e1bcbc35a536abb93ec88dc3bfc288b23bf/3.2.0/enterprise/Dockerfile)
+-	[`3.1.4`, `3.1` (*3.1.4/community/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/86c2eb04e9f44e2c6febf5c14d0a7fe67c81f35c/3.1.4/community/Dockerfile)
+-	[`3.1.4-enterprise`, `3.1-enterprise` (*3.1.4/enterprise/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/86c2eb04e9f44e2c6febf5c14d0a7fe67c81f35c/3.1.4/enterprise/Dockerfile)
 -	[`3.1.3` (*3.1.3/community/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/a50bed8c92cb9d24adb6b5a7353455c222b1be9d/3.1.3/community/Dockerfile)
 -	[`3.1.3-enterprise` (*3.1.3/enterprise/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/a50bed8c92cb9d24adb6b5a7353455c222b1be9d/3.1.3/enterprise/Dockerfile)
 -	[`3.1.2` (*3.1.2/community/Dockerfile*)](https://github.com/neo4j/docker-neo4j-publish/blob/eacdd603aae3c521ac67728a54a7c5609d5581eb/3.1.2/community/Dockerfile)

+ 6 - 5
ros/README.md

@@ -89,11 +89,12 @@ With the advancements and standardization of software containers, roboticists ar
 
 ## Deployment suggestions
 
-The available tags include supported distros along with a hierarchy tags based off the most common meta-package dependencies, designed to have a small footprint and simple configuration:  
-- `ros-core`: barebone ROS install  
-- `ros-base`: basic tools and libraries (also tagged with distro name with LTS version as `latest`)  
-- `robot`: basic install for robots  
-- `perception`: basic install for perception tasks
+The available tags include supported distros along with a hierarchy tags based off the most common meta-package dependencies, designed to have a small footprint and simple configuration:
+
+-	`ros-core`: barebone ROS install
+-	`ros-base`: basic tools and libraries (also tagged with distro name with LTS version as `latest`)
+-	`robot`: basic install for robots
+-	`perception`: basic install for perception tasks
 
 The rest of the common meta-packages such as `desktop` and `desktop-full` are hosted on automatic build repos under OSRF's Docker Hub profile [here](https://hub.docker.com/r/osrf/ros/). These meta-packages include graphical dependencies and hook a host of other large packages such as X11, X server, etc. So in the interest of keep the official images lean and secure, the desktop packages are just be hosted with OSRF's profile.
 

+ 73 - 53
silverpeas/README.md

@@ -79,10 +79,12 @@ These environment variables can be also defined as properties into the Silverpea
 
 In [Docker Hub](https://hub.docker.com/), no Docker images of Microsoft SQLServer are currently available, but you will find a lot of images of PostgreSQL. For example, with an [official PostgreSQL docker image](https://hub.docker.com/_/postgres/), you can start a PostgreSQL instance initialized with a superuser `postgres` with as password `mysecretpassword`:
 
-	$ docker run --name postgresql -d \
-	    -e POSTGRES_PASSWORD="mysecretpassword" \
-	    -v postgresql-data:/var/lib/postgresql/data \
-	    postgres:9.6
+```console
+$ docker run --name postgresql -d \
+    -e POSTGRES_PASSWORD="mysecretpassword" \
+    -v postgresql-data:/var/lib/postgresql/data \
+    postgres:9.6
+```
 
 We recommend strongly to mount the directory with the database file on the host so the data won't be lost when upgrading PostgreSQL to a newer version (a Data Volume Container can be used instead). For any information how to start a PostgreSQL container, you can refer its [documentation]((https://hub.docker.com/_/postgres/).
 
@@ -92,14 +94,16 @@ Once the database system is running, a database for Silverpeas has to be created
 
 Finally, a Silverpeas instance can be started by specifying the required database access parameters with the environment variables. In the example, the database is named `Silverpeas` and the priviledged user is `silverpeas` with as password `thesilverpeaspassword`:
 
-	$ docker run --name silverpeas -p 8080:8000 -d \
-	    -e DB_NAME="Silverpeas" \
-	    -e DB_USER="silverpeas" \
-	    -e DB_PASSWORD="thesilverpeaspassword" \
-	    -v silverpeas-log:/opt/silverpeas/log \
-	    -v silverpeas-data:/opt/silverpeas/data \
-	    --link postgresql:database \
-	    silverpeas
+```console
+$ docker run --name silverpeas -p 8080:8000 -d \
+    -e DB_NAME="Silverpeas" \
+    -e DB_USER="silverpeas" \
+    -e DB_PASSWORD="thesilverpeaspassword" \
+    -v silverpeas-log:/opt/silverpeas/log \
+    -v silverpeas-data:/opt/silverpeas/data \
+    --link postgresql:database \
+    silverpeas
+```
 
 Here, as the PostgreSQL database is linked under the alias `database`, we don't have to explicitly indicate its hostname with the `DB_SERVER` environment variable. The Silverpeas images expose the 8000 port and here this port is mapped to the 8080 port of the host; Silverpeas is then accessible at `http://localhost:8080/silverpeas`. You can sign in Silverpeas with the administrator account `SilverAdmin` and with as password `SilverAdmin`.
 
@@ -109,12 +113,14 @@ By default, some volumes are created inside the container, so that we can access
 
 The Silverpeas global configuration is defined in the `/opt/silverpeas/configuration/config.properties` file whose a sample can be found [here](https://raw.githubusercontent.com/Silverpeas/Silverpeas-Distribution/master/src/main/dist/configuration/sample_config.properties) or in the container directory `/opt/silverpeas/configuration/`. You can explicitly create the `config.properties` file with, additionally to the database access parameters, your peculiar configuration parameters and then start a Silverpeas instance with this configuration file:
 
-	$ docker run --name silverpeas -p 8080:8000 -d \
-	    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/config.properties
-	    -v silverpeas-log:/opt/silverpeas/log \
-	    -v silverpeas-data:/opt/silverpeas/data \
-	    --link postgresql:database \
-	    silverpeas
+```console
+$ docker run --name silverpeas -p 8080:8000 -d \
+    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/config.properties
+    -v silverpeas-log:/opt/silverpeas/log \
+    -v silverpeas-data:/opt/silverpeas/data \
+    --link postgresql:database \
+    silverpeas
+```
 
 where `/etc/silverpeas/config.properties` is your own configuration file on the host. For security reason, we strongly recommend to set explicitly the administrator's credentials with the properties `SILVERPEAS_ADMIN_LOGIN` and `SILVERPEAS_ADMIN_PASSWORD` in the `config.properties` file. (Don't forget to set also the administrator email address with the property `SILVERPEAS_ADMIN_EMAIL`.)
 
@@ -122,12 +128,14 @@ where `/etc/silverpeas/config.properties` is your own configuration file on the
 
 For a database system running on the host (or on a remote host) with 192.168.1.14 as IP address, you have to specify this host both to the container at starting and to Silverpeas by defining it into its global configuration file:
 
-	$ docker run --name silverpeas -p 8080:8000 -d \
-	    --add-host=database:192.168.1.14 \
-	    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/config.properties \
-	    -v silverpeas-log:/opt/silverpeas/log \
-	    -v silverpeas-data:/opt/silverpeas/data \
-	    silverpeas
+```console
+$ docker run --name silverpeas -p 8080:8000 -d \
+    --add-host=database:192.168.1.14 \
+    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/config.properties \
+    -v silverpeas-log:/opt/silverpeas/log \
+    -v silverpeas-data:/opt/silverpeas/data \
+    silverpeas
+```
 
 where `database` is the hostname referred by the `DB_SERVER` parameter in your `/etc/silverpeas/config.properties` file as the host running the database system and that is mapped here to the actual IP address of this host. The hostname is added in the `/etc/hosts` file in the container.
 
@@ -161,32 +169,38 @@ All these different kind of data have to be consistent for a given state of Silv
 
 To define a Data Volume Container for Silverpeas, for example:
 
-	$ docker create --name silverpeas-store \
-	    -v silverpeas-data:/opt/silverpeas/data \
-	    -v silverpeas-log:/opt/silverpeas/log \
-	    -v silverpeas-workflows:/opt/silverpeas/xmlcomponents/workflows \
-	    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/properties \
-	    silverpeas \
-	    /bin/true
+```console
+$ docker create --name silverpeas-store \
+    -v silverpeas-data:/opt/silverpeas/data \
+    -v silverpeas-log:/opt/silverpeas/log \
+    -v silverpeas-workflows:/opt/silverpeas/xmlcomponents/workflows \
+    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/properties \
+    silverpeas \
+    /bin/true
+```
 
 Then to mount the volumes in the Silverpeas container:
 
-	$ docker run --name silverpeas -p 8080:8000 -d \
-	    --link postgresql:database \
-	    --volumes-from silverpeas-store \
-	    silverpeas
+```console
+$ docker run --name silverpeas -p 8080:8000 -d \
+    --link postgresql:database \
+    --volumes-from silverpeas-store \
+    silverpeas
+```
 
 If you have to customize the settings of Silverpeas or add, for example, a new database definition, then specify these settings with the Data Volume Container, so that they will be available to the next versions of Silverpeas which will be then configured correctly like your previous Silverpeas installation:
 
-	$ docker create --name silverpeas-store \
-	    -v silverpeas-data:/opt/silverpeas/data \
-	    -v silverpeas-log:/opt/silverpeas/log \
-	    -v silverpeas-properties:/opt/silverpeas/properties \
-	    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/properties \
-	    -v /etc/silverpeas/CustomerSettings.xml:/opt/silverpeas/configuration/silverpeas/CustomerSettings.xml \
-	    -v /etc/silverpeas/my-datasource.cli:/opt/silverpeas/configuration/jboss/my-datasource.cli \
-	    silverpeas \
-	    /bin/true
+```console
+$ docker create --name silverpeas-store \
+    -v silverpeas-data:/opt/silverpeas/data \
+    -v silverpeas-log:/opt/silverpeas/log \
+    -v silverpeas-properties:/opt/silverpeas/properties \
+    -v /etc/silverpeas/config.properties:/opt/silverpeas/configuration/properties \
+    -v /etc/silverpeas/CustomerSettings.xml:/opt/silverpeas/configuration/silverpeas/CustomerSettings.xml \
+    -v /etc/silverpeas/my-datasource.cli:/opt/silverpeas/configuration/jboss/my-datasource.cli \
+    silverpeas \
+    /bin/true
+```
 
 # Document conversion
 
@@ -194,9 +208,11 @@ Some features in Silverpeas (export, preview, content visualization, ...) requir
 
 Once a Data Volume Container is created for Silverpeas as explained in the section above, you have to link it with the Docker image running LibreOffice as a daemon in order the program have access the documents to convert:
 
-	$ docker run --name libreoffice -d \
-	    --volumes-from silverpeas-store \
-	    xcgd/libreoffice
+```console
+$ docker run --name libreoffice -d \
+    --volumes-from silverpeas-store \
+    xcgd/libreoffice
+```
 
 Check the port at which the LibreOffice image is listening and then defines it in the Silverpeas configuration. In our example, `xcgd/libreoffice` listens by default the port 8997. The configuration parameters to communicate with LibreOffice are defined by the two following properties:
 
@@ -210,11 +226,13 @@ These properties have to be defined in the Silverpeas global configuration file
 
 Then the Docker image of Silverpeas can be ran:
 
-	$ docker run --name silverpeas -p 8080:8000 -d \
-	    --link postgresql:database \
-	    --link libreoffice:libreoffice \
-	    --volumes-from silverpeas-store \
-	    silverpeas
+```console
+$ docker run --name silverpeas -p 8080:8000 -d \
+    --link postgresql:database \
+    --link libreoffice:libreoffice \
+    --volumes-from silverpeas-store \
+    silverpeas
+```
 
 # Logs
 
@@ -222,7 +240,9 @@ You can follow the activity of Silverpeas by watching the logs generated in the
 
 The output of Wildfly is redirected into the container standard output and so it can be watched as following:
 
-	$ docker logs -f silverpeas
+```console
+$ docker logs -f silverpeas
+```
 
 # License
 

+ 27 - 9
storm/README.md

@@ -60,29 +60,41 @@ Apache Storm is a distributed computation framework written predominantly in the
 
 Assuming you have `topology.jar` in the current directory.
 
-	$ docker run -it -v $(pwd)/topology.jar:/topology.jar storm storm jar /topology.jar org.apache.storm.starter.ExclamationTopology
+```console
+$ docker run -it -v $(pwd)/topology.jar:/topology.jar storm storm jar /topology.jar org.apache.storm.starter.ExclamationTopology
+```
 
 ## Setting up a minimal Storm cluster
 
 1.	[Apache Zookeeper](https://zookeeper.apache.org/) is a must for running a Storm cluster. Start it first. Since the Zookeeper "fails fast" it's better to always restart it.
 
-		$ docker run -d --restart always --name some-zookeeper zookeeper
+	```console
+	$ docker run -d --restart always --name some-zookeeper zookeeper
+	```
 
 2.	The Nimbus daemon has to be connected with the Zookeeper. It's also a "fail fast" system.
 
-		$ docker run -d --restart always --name some-nimbus --link some-zookeeper:zookeeper storm storm nimbus
+	```console
+	$ docker run -d --restart always --name some-nimbus --link some-zookeeper:zookeeper storm storm nimbus
+	```
 
 3.	Finally start a single Supervisor node. It will talk to the Nimbus and Zookeeper.
 
-		$ docker run -d --restart always --name supervisor --link some-zookeeper:zookeeper --link some-nimbus:nimbus storm storm supervisor
+	```console
+	$ docker run -d --restart always --name supervisor --link some-zookeeper:zookeeper --link some-nimbus:nimbus storm storm supervisor
+	```
 
 4.	Now you can submit a topology to our cluster.
 
-		$ docker run --link some-nimbus:nimbus -it --rm -v $(pwd)/topology.jar:/topology.jar storm storm jar /topology.jar org.apache.storm.starter.WordCountTopology topology
+	```console
+	$ docker run --link some-nimbus:nimbus -it --rm -v $(pwd)/topology.jar:/topology.jar storm storm jar /topology.jar org.apache.storm.starter.WordCountTopology topology
+	```
 
 5.	Optionally, you can start the Storm UI.
 
-		$ docker run -d -p 8080:8080 --restart always --name ui --link some-nimbus:nimbus storm storm ui
+	```console
+	$ docker run -d -p 8080:8080 --restart always --name ui --link some-nimbus:nimbus storm storm ui
+	```
 
 ## ... via [`docker-compose`](https://github.com/docker/compose)
 
@@ -129,11 +141,15 @@ This image uses [default configuration](https://github.com/apache/storm/blob/v1.
 
 1.	Using command line arguments.
 
-		$ docker run -d --restart always --name nimbus storm storm nimbus -c storm.zookeeper.servers='["zookeeper"]'
+	```console
+	$ docker run -d --restart always --name nimbus storm storm nimbus -c storm.zookeeper.servers='["zookeeper"]'
+	```
 
 2.	Assuming you have `storm.yaml` in the current directory you can mount it as a volume.
 
-		$ docker run -it -v $(pwd)/storm.yaml:/conf/storm.yaml storm storm nimbus
+	```console
+	$ docker run -it -v $(pwd)/storm.yaml:/conf/storm.yaml storm storm nimbus
+	```
 
 ## Logging
 
@@ -143,7 +159,9 @@ This image uses [default logging configuration](https://github.com/apache/storm/
 
 No data are persisted by default. For convenience there are `/data` and `/logs` directories in the image owned by `storm` user. Use them accordingly to persist data and logs using volumes.
 
-	$ docker run -it -v /logs -v /data storm storm nimbus
+```console
+$ docker run -it -v /logs -v /data storm storm nimbus
+```
 
 *Please be noticed that using paths other than those predefined is likely to cause permission denied errors. It's because for [security reasons](https://docs.docker.com/engine/userguide/eng-image/dockerfile_best-practices/#user) the Storm is running under the non-root `storm` user.*
 

+ 15 - 5
zookeeper/README.md

@@ -56,17 +56,23 @@ Apache ZooKeeper is a software project of the Apache Software Foundation, provid
 
 ## Start a Zookeeper server instance
 
-	$ docker run --name some-zookeeper --restart always -d zookeeper
+```console
+$ docker run --name some-zookeeper --restart always -d zookeeper
+```
 
 This image includes `EXPOSE 2181 2888 3888` (the zookeeper client port, follower port, election port respectively), so standard container linking will make it automatically available to the linked containers. Since the Zookeeper "fails fast" it's better to always restart it.
 
 ## Connect to Zookeeper from an application in another Docker container
 
-	$ docker run --name some-app --link some-zookeeper:zookeeper -d application-that-uses-zookeeper
+```console
+$ docker run --name some-app --link some-zookeeper:zookeeper -d application-that-uses-zookeeper
+```
 
 ## Connect to Zookeeper from the Zookeeper command line client
 
-	$ docker run -it --rm --link some-zookeeper:zookeeper zookeeper zkCli.sh -server zookeeper
+```console
+$ docker run -it --rm --link some-zookeeper:zookeeper zookeeper zkCli.sh -server zookeeper
+```
 
 ## ... via [`docker-compose`](https://github.com/docker/compose)
 
@@ -113,13 +119,17 @@ Consider using [Docker Swarm](https://www.docker.com/products/docker-swarm) when
 
 Zookeeper configuration is located in `/conf`. One way to change it is mounting your config file as a volume:
 
-	$ docker run --name some-zookeeper --restart always -d -v $(pwd)/zoo.cfg:/conf/zoo.cfg zookeeper
+```console
+$ docker run --name some-zookeeper --restart always -d -v $(pwd)/zoo.cfg:/conf/zoo.cfg zookeeper
+```
 
 ## Environment variables
 
 ZooKeeper recommended defaults are used if `zoo.cfg` file is not provided. They can be overridden using the following environment variables.
 
-	$ docker run -e "ZOO_INIT_LIMIT=10" --name some-zookeeper --restart always -d 31z4/zookeeper
+```console
+$ docker run -e "ZOO_INIT_LIMIT=10" --name some-zookeeper --restart always -d 31z4/zookeeper
+```
 
 ### `ZOO_TICK_TIME`