Since Elasticsearch is exposed via an HTTP API, we can user our nginx server to proxy Elasticsearch requests using the HTTPS protocol.
Let’s say you have your local dev environment configured to use SSL. Your dev site is accessible at https://mysite.dev/
. Wonderful! Now you need to add Elasticsearch to your project. Let’s add it to docker-compose.yml
, something like:
version: "2"
services:
elasticsearch:
image: docker.elastic.co/elasticsearch/elasticsearch:5.2.2
environment:
- xpack.security.enabled=false
- bootstrap.memory_lock=true
- "ES_JAVA_OPTS=-Xms512m -Xmx512m"
ulimits:
memlock:
soft: -1
hard: -1
nofile:
soft: 65536
hard: 65536
mem_limit: 1g
volumes:
- elasticsearchindex:/usr/share/elasticsearch/data
ports:
- "9200"
network_mode: "bridge"
# and some other services like PHP, nginx, memcached, mysql
volumes:
elasticsearchindex:
How do you make requests to Elasticsearch from the browser?
Option 1: Set up a proxy in your app. This probably resembles what you’ll ultimately get in production. You don’t really need any security on Elasticsearch for local dev, but in production it will need some sort of access control so users can’t send arbitrary requests to the server. If you’re not using a third-party service that already handles this for you, this is where you’ll filter out invalid or dangerous requests. I prefer to let more experienced hands manage server security for me, though, and this is a lot of overhead just to set up a local dev server.
Option 2: Expose Elasticsearch directly. Since I don’t need security locally, I could just open up port 9200 on my container and make requests directly to it from the browser at http://localhost:9200/
. Notice the protocol there, though. If my local site is at https://mysite.dev/
, then the browser will block insecure requests to Elasticsearch.
Option 3: Use nginx as a proxy. I’m already using a reverse proxy in front of my project containers. It terminates the SSL connections and then passes through unencrypted requests to each project’s nginx server. The project’s nginx container doesn’t need to deal with SSL. It listens on port 80 and passes requests to PHP with fastcgi.
server {
listen 80 default_server;
server_name mysite.dev;
# ... more server boilerplate
}
Since Elasticsearch is exposed via an HTTP API, we can create another server
block to proxy Elasticsearch requests. First, make sure the nginx container can talk to the Elasticsearch container. In docker-compose.yml
:
nginx:
image: nginx:stable-alpine
environment:
- VIRTUAL_HOST=mysite.dev,*.mysite.dev
volumes:
- ./nginx/default.conf:/etc/nginx/conf.d/default.conf:ro
- ./nginx/elasticsearch-proxy.conf:/etc/nginx/conf.d/elasticsearch-proxy.conf:ro
- ./nginx/php.conf:/etc/nginx/php.conf:ro
links:
- php
- elasticsearch
ports:
- "80"
network_mode: "bridge"
And then create elasticsearch-proxy.conf
to handle the requests:
upstream es {
server elasticsearch:9200;
keepalive 15;
}
server {
listen 80;
server_name search.mysite.dev;
location / {
proxy_pass http://es;
proxy_http_version 1.1;
proxy_set_header Connection "Keep-Alive";
proxy_set_header Proxy-Connection "Keep-Alive";
}
}
Now we can make requests to Elasticsearch from the browser at https://search.mysite.dev/
. The nginx proxy will handle the SSL termination, and communicate with Elasticsearch using its standard HTTP API.