diff --git a/CHANGELOG.md b/CHANGELOG.md
index 4467ece33e62e9f29dac413c1828847ed9b766c3..c61fef2697c6fe4741982d8a4730dac53dc28a93 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,3 +1,9 @@
+## new in jupyter-openbis-extension 0.5.0
+
+- splitted project into jupyter-openbis-server and jupyter-openbis-extension
+- jupyter-openbis-server is the Python-part which talks to both openBIS and the Jupyter extension
+- jupyter-opnebis-server is also used by the jupyterlab-openbis notebook extension
+
 ## new in jupyter-openbis-extension 0.4.0
 
 - made it compatible to pybis-1.9.x
diff --git a/MANIFEST.in b/MANIFEST.in
index d0f16ab5ae0d656e219f4758e6f9470047f1dcf0..df7c4348a3ee680361930d00aaeb29499d4f33e0 100644
--- a/MANIFEST.in
+++ b/MANIFEST.in
@@ -5,7 +5,7 @@ include *.md
 include LICENSE
 
 # Include all files
-recursive-include jupyter-openbis-extension *.py *.js
+recursive-include jupyter-openbis-extension *.js
 
 recursive-include jupyter-config *.json
 recursive-include jupyter-openbis-extension/static *.js
diff --git a/jupyter-config/jupyter_notebook_config.d/jupyter_openbis_extension.json b/jupyter-config/jupyter_notebook_config.d/jupyter_openbis_extension.json
deleted file mode 100644
index 501e38c9704dcd0325c7fa818d07a68d51e1873f..0000000000000000000000000000000000000000
--- a/jupyter-config/jupyter_notebook_config.d/jupyter_openbis_extension.json
+++ /dev/null
@@ -1,7 +0,0 @@
-{
-  "NotebookApp": {
-    "nbserver_extensions": {
-      "jupyter-openbis-extension.server": true
-    }
-  }
-}
diff --git a/jupyter-openbis-extension/__init__.py b/jupyter-openbis-extension/__init__.py
deleted file mode 100644
index c4c2f3ea1bd3b4b9375e676f33f3c6fd73ebcab2..0000000000000000000000000000000000000000
--- a/jupyter-openbis-extension/__init__.py
+++ /dev/null
@@ -1,26 +0,0 @@
-name = 'jupyter-openbis-extension.server'
-__author__ = 'Swen Vermeul'
-__email__ = 'swen@ethz.ch'
-__version__ = '0.4.0'
-
-def _jupyter_server_extension_paths():
-    return [{
-        "module": "jupyter-openbis-extension.server"
-    }]
-
-# Jupyter Extension points
-def _jupyter_nbextension_paths():
-    return [{
-        'name': 'jupyter-openbis-extension',
-        'label': 'Jupyter openBIS extension',
-        'section': "notebook",
-    #    # the path relative to the `jupyter-openbis-extension` directory containing the JavaScript
-        'src': "static",
-    #    # directory in the `nbextension/` namespace
-        'dest': "openbis",
-        # _also_ in the `nbextension/` namespace
-        'require' : "openbis/main"
-    }]
-
-def load_jupyter_server_extension(nbapp):
-    nbapp.log.info("jupyter-openbis-extension module enabled!")
diff --git a/jupyter-openbis-extension/connection.py b/jupyter-openbis-extension/connection.py
deleted file mode 100644
index 557bc4966487975ee684799b4d2e2473fe4264ac..0000000000000000000000000000000000000000
--- a/jupyter-openbis-extension/connection.py
+++ /dev/null
@@ -1,191 +0,0 @@
-import os
-from pybis import Openbis
-from notebook.base.handlers import IPythonHandler
-
-openbis_connections = {}
-
-def register_connection(connection_info):
-
-    conn = OpenBISConnection(
-        name                = connection_info.get('name'),
-        url                 = connection_info.get('url'),
-        verify_certificates = connection_info.get('verify_certificates', False),
-        username            = connection_info.get('username'),
-        password            = connection_info.get('password'),
-        http_only           = connection_info.get('http_only', False),
-        status              = 'not connected',
-    )
-    openbis_connections[conn.name] = conn
-    return conn
-
-
-class OpenBISConnection:
-    """register an openBIS connection
-    """
-
-    def __init__(self, **kwargs):
-        for needed_key in ['name', 'url']:
-            if needed_key not in kwargs:
-                raise KeyError("{} is missing".format(needed_key))
-
-        for key in kwargs:
-            setattr(self, key, kwargs[key])
-
-        openbis = Openbis(
-            url = self.url,
-            verify_certificates = self.verify_certificates,
-            allow_http_but_do_not_use_this_in_production_and_only_within_safe_networks = self.http_only
-        )
-        self.openbis = openbis
-        self.status = "not connected"
-
-    def is_session_active(self):
-        return self.openbis.is_session_active()
-
-    def check_status(self):
-        if self.openbis.is_session_active():
-            self.status = "connected"
-        else:
-            self.status = "not connected"
-
-    def login(self, username=None, password=None):
-        if username is None:
-            username=self.username
-        if password is None:
-            password=self.password
-        self.openbis.login(
-            username = username,
-            password = password
-        )
-        # store username and password in memory
-        self.username = username
-        self.password = password
-        self.status  = 'connected'
-
-    def get_info(self):
-        return {
-            'name'    : self.name,
-            'url'     : self.url,
-            'status'  : self.status,
-            'username': self.username,
-            'password': self.password,
-        }
-
-class OpenBISConnections(IPythonHandler):
-
-    def _notebook_dir(self):
-        notebook_dir = os.getcwd()
-        if 'SingleUserNotebookApp' in self.config and 'notebook_dir' in self.config.SingleUserNotebookApp:
-            notebook_dir = self.config.SingleUserNotebookApp.notebook_dir
-        elif 'notebook_dir' in self.config.NotebookApp:
-            notebook_dir = self.config.NotebookApp.notebook_dir
-        return notebook_dir
-
-    def post(self):
-        """create a new connection
-
-        :return: a new connection object
-        """
-        data = self.get_json_body()
-        conn = register_connection(data)
-        if conn.username and conn.password:
-            try:
-                conn.login()
-            except Exception:
-                pass
-        self.get()
-        return
-
-    def get(self):
-        """returns all available openBIS connections
-        """
-
-        connections= []
-        for conn in openbis_connections.values():
-            conn.check_status()
-            connections.append(conn.get_info())
-
-        self.write({
-            'status'       : 200,
-            'connections'  : connections,
-            'notebook_dir' : self._notebook_dir()
-        })
-        return
-
-
-class OpenBISConnectionHandler(IPythonHandler):
-    """Handle the requests to /openbis/conn
-    """
-
-    def _notebook_dir(self):
-        notebook_dir = os.getcwd()
-        if 'SingleUserNotebookApp' in self.config and 'notebook_dir' in self.config.SingleUserNotebookApp:
-            notebook_dir = self.config.SingleUserNotebookApp.notebook_dir
-        elif 'notebook_dir' in self.config.NotebookApp:
-            notebook_dir = self.config.NotebookApp.notebook_dir
-        return notebook_dir
-
-    def put(self, connection_name):
-        """reconnect to a current connection
-        :return: an updated connection object
-        """
-        data = self.get_json_body()
-
-        try:
-            conn = openbis_connections[connection_name]
-        except KeyError:
-            self.set_status(404)
-            self.write({
-                "reason" : 'No such connection: {}'.format(data)
-            })
-            return
-
-        try:
-            conn.login(data.get('username'), data.get('password'))
-        except ConnectionError:
-            self.set_status(500)
-            self.write({
-                "reason": "Could not establish connection to {}".format(connection_name)
-            })
-            return
-        except ValueError:
-            self.set_status(401)
-            self.write({
-                "reason": "Incorrect username or password for {}".format(connection_name)
-            })
-            return
-        except Exception:
-            self.set_status(500)
-            self.write({
-                "reason": "General Network Error"
-            })
-
-        self.write({
-            'status'     : 200,
-            'connection' : conn.get_info(),
-            ''           : self._notebook_dir()
-        })
-
-    def get(self, connection_name):
-        """returns  information about a connection name
-        """
-
-        try:
-            conn = openbis_connections[connection_name]
-        except KeyError:
-            self.set_status(404)
-            self.write({
-                "reason" : 'No such connection: {}'.format(connection_name)
-            })
-            return
-
-        conn.check_status()
-
-        self.write({
-            'status'        : 200,
-            'connection'    : conn.get_info(),
-            'noteboook_dir' : self._notebook_dir()
-        })
-        return
-
-
diff --git a/jupyter-openbis-extension/dataset.py b/jupyter-openbis-extension/dataset.py
deleted file mode 100644
index 884c519c1a5159ab260f73208b55dae2dc3e0069..0000000000000000000000000000000000000000
--- a/jupyter-openbis-extension/dataset.py
+++ /dev/null
@@ -1,288 +0,0 @@
-import os
-from urllib.parse import unquote
-from notebook.base.handlers import IPythonHandler
-from .connection import openbis_connections
-
-class DataSetDownloadHandler(IPythonHandler):
-    """Handle the requests for /openbis/dataset/connection/permId"""
-
-    def download_data(self, conn, permId, downloadPath=None):
-        if not conn.is_session_active():
-            try:
-                conn.login()
-            except Exception as exc:
-                self.set_status(500)
-                self.write({
-                    "reason" : 'connection to {} could not be established: {}'.format(conn.name, exc)
-                })
-                return
-
-        try:
-            dataset = conn.openbis.get_dataset(permId)
-        except Exception as exc:
-            self.set_status(404)
-            self.write({
-                "reason" : 'No such dataSet found: {}'.format(permId)
-            })
-            return
-
-        # dataset was found, download the data to the disk
-        try:
-            destination = dataset.download(destination=downloadPath)
-        except Exception as exc:
-            self.set_status(500)
-            self.write({
-                "reason": 'Data for DataSet {} could not be downloaded: {}'.format(permId, exc)
-            })
-            return
-
-        # return success message
-        path = os.path.join(downloadPath, dataset.permId)
-        self.write({
-            'url'       : conn.url,
-            'permId'    : dataset.permId,
-            'path'      : path,
-            'dataStore' : dataset.dataStore,
-            'location'  : dataset.physicalData.location,
-            'size'      : dataset.physicalData.size,
-            'files'     : dataset.file_list,
-            'statusText': 'Data for DataSet {} was successfully downloaded to: {}.'.format(dataset.permId, path)
-        })
-
-    def get(self, **params):
-        """Handle a request to /openbis/dataset/connection_name/permId
-        download the data and return a message
-        """
-
-        try:
-            conn = openbis_connections[params['connection_name']]
-        except KeyError:
-            self.set_status(404)
-            self.write({
-                "reason":'connection {} was not found'.format(params['connection_name'])
-            })
-            return
-
-        results = self.download_data(conn=conn, permId=params['permId'], downloadPath=params['downloadPath'])
-
-
-class DataSetTypesHandler(IPythonHandler):
-    def get(self, **params):
-        """Handle a request to /openbis/datasetTypes/connection_name
-        This meta-metadata is used in the dataset upload dialog (uploadDialog.js)
-        to check data directly in the UI
-
-        Returns all datasetTypes of a given connection
-        - with all assigned properties
-        - with some details about the property types
-        - with the vocabulary, if exists
-
-        The result will be cached, as it is a costly operation with many fetches
-        """
-
-        try:
-            conn = openbis_connections[params['connection_name']]
-        except KeyError:
-            self.set_status(404)
-            self.write({
-                "reason":'connection {} was not found'.format(params['connection_name'])
-            })
-            return
-
-        if getattr(conn, 'dataset_types', False):
-            self.write({
-                "dataSetTypes": conn.dataset_types
-            })
-            return
-
-        try:
-            dataset_types = conn.openbis.get_dataset_types()
-
-            # get all dataset types
-            ds_type_dicts = []
-            for dt in conn.openbis.get_dataset_types():
-                dt_dict = dt.attrs.all()
-                # get property assignments for every dataset-type
-                # and add them in the key «propertyAssignments»
-                pas = dt.get_property_assignments()
-                pa_dicts = pas.df[['propertyType','mandatory','ordinal','section']].to_dict(orient='records')
-                dt_dict['propertyAssignments'] = pa_dicts
-
-                for pa_dict in pa_dicts:
-                    # add a few more attributes to the property assignments
-                    pt = conn.openbis.get_property_type(pa_dict['propertyType'])
-                    pa_dict['code'] = pt.code
-                    pa_dict['label'] = pt.label
-                    pa_dict['description'] = pt.description
-                    pa_dict['dataType'] = pt.dataType
-                    # add vocabulary, if exists, as key «terms»
-                    if pt.dataType == 'CONTROLLEDVOCABULARY':
-                        terms = conn.openbis.get_terms(pt.vocabulary)
-                        terms_dict = terms.df[['code','label','description','official','ordinal']].to_dict(orient='records')
-                        pa_dict['terms'] = terms_dict
-
-                ds_type_dicts.append(dt_dict)
-
-            self.write({
-                "dataSetTypes": ds_type_dicts
-            })
-            conn.dataset_types = ds_type_dicts
-            return
-
-        except Exception as e:
-            print(e)
-            self.set_status(500)
-            self.write({
-                "reason":'Could not fetch dataset-types: {}'.format(e)
-            })
-            return
-
-
-class DataSetUploadHandler(IPythonHandler):
-    """Handle the POST requests for /openbis/dataset/connection_name"""
-
-    def _notebook_dir(self):
-        notebook_dir = os.getcwd()
-        if 'SingleUserNotebookApp' in self.config and 'notebook_dir' in self.config.SingleUserNotebookApp:
-            notebook_dir = self.config.SingleUserNotebookApp.notebook_dir
-        elif 'notebook_dir' in self.config.NotebookApp:
-            notebook_dir = self.config.NotebookApp.notebook_dir
-        return notebook_dir
-
-    def upload_data(self, conn, data):
-        if not conn.is_session_active():
-            try:
-                conn.login()
-            except Exception as e:
-                print(e)
-                self.set_status(500)
-                self.write({
-                    "reason": 'connection to {} could not be established: {}'.format(conn.name, e)
-                })
-                return
-
-        errors = []
-
-        sample = None
-        experiment = None
-
-        if (data.get('entityIdentifier')):
-            sample = None
-            experiment = None
-            try:
-                sample = conn.openbis.get_sample(data.get('entityIdentifier'))
-            except Exception as e:
-                pass
-            if sample is None:
-                try:
-                    experiment = conn.openbis.get_experiment(data.get('entityIdentifier'))
-                except Exception as e:
-                    pass
-
-            if sample is None and experiment is None:
-                errors.append(
-                    {"entityIdentifier" : 'No such sample or experiment: {}'.format(data.get('entityIdentifier')) }
-                )
-        else:
-            errors.append(
-                {"entityIdentifier": "please provide a sample or experiment identifier"}
-            )
-
-        parents = []
-        if data.get('parents'):
-            parents = data.get('parents')
-            for parent in parents:
-                try:
-                    conn.openbis.get_dataset(parent)
-                except Exception as e:
-                    errors.append({
-                        "parent": "Parent DataSet not found: {}".format(parent)
-                    })
-
-        filenames = []
-        notebook_dir = self._notebook_dir()
-        for filename in data.get('files'):
-            filename = unquote(filename)
-            full_filename_path = os.path.join(notebook_dir, filename)
-            if os.path.isfile(full_filename_path):
-                filenames.append(full_filename_path)
-            else:
-                errors.append({
-                    "file": "File not found: {}".format(full_filename_path)
-                })
-
-        try:
-            dataset = conn.openbis.new_dataset(
-                type   = data.get('type'),
-                sample = sample,
-                parents = parents,
-                experiment = experiment,
-                files  = filenames,
-            )
-        except Exception as e:
-            self.set_status(500)
-            self.write({
-                "reason": 'Error while creating the dataset: {}'.format(e)
-            })
-
-        # try to set the properties
-        if data.get('props'):
-            props = data.get('props')
-            for prop, value in props.items():
-                try:
-                    setattr(dataset.props, prop.lower(), value)
-                except Exception as e:
-                    errors.append({
-                        "prop."+prop : str(e)
-                    })
-
-        # check if any mandatory property is missing
-        for prop_name, prop in dataset.props._property_names.items():
-            if prop['mandatory']:
-                if getattr(dataset.props, prop_name) is None or getattr(dataset.props, prop_name) == "":
-                    errors.append({
-                        "prop."+prop_name : "is mandatory"
-                    })
-
-        # write errors back if already occured
-        if errors:
-            self.set_status(500)
-            self.write({ "errors": errors })
-            return
-
-        try:
-            dataset.save()
-        except Exception as e:
-            errors.append({
-                "save": 'Error while saving the dataset: {}'.format(e)
-            })
-
-        # write errors back if they occured
-        if errors:
-            self.set_status(500)
-            self.write({ "errors": errors })
-        else:
-            # ...or return a success message
-            self.write({
-                'status': 200,
-                'statusText': 'Jupyter Notebook was successfully uploaded to: {} with permId: {}'.format(conn.name, dataset.permId)
-            })
-        print('Jupyter Notebook was successfully uploaded to: {} with permId: {}'.format(conn.name, dataset.permId))
-
-    def post(self, **params):
-        """Handle a request to /openbis/dataset/connection_name/permId
-        download the data and return a message
-        """
-
-        try:
-            conn = openbis_connections[params['connection_name']]
-        except KeyError:
-            self.write({
-                "reason": 'connection {} was not found'.format(params['connection_name'])
-            })
-            return
-
-        data = self.get_json_body()
-        self.upload_data(conn=conn,data=data)
-
-
diff --git a/jupyter-openbis-extension/sample.py b/jupyter-openbis-extension/sample.py
deleted file mode 100644
index 0da1985ca6e6d1091177b34c2555c4f5349eca21..0000000000000000000000000000000000000000
--- a/jupyter-openbis-extension/sample.py
+++ /dev/null
@@ -1,89 +0,0 @@
-from notebook.base.handlers import IPythonHandler
-import numpy as np
-import os
-from .connection import openbis_connections
-from urllib.parse import parse_qs
-
-
-def get_entity_for_identifier(conn, identifier):
-    entity = None
-    try:
-        entity = conn.openbis.get_sample(identifier)
-    except Exception as exc:
-        pass
-
-    if entity is None:
-        try:
-            entity = conn.openbis.get_experiment(identifier)
-        except Exception as exc:
-            pass
-    return entity
-
-
-def get_datasets(entity, start_with=None, count=None):
-
-    datasets = entity.get_datasets(start_with=start_with, count=count)
-    totalCount = datasets.totalCount
-    df = datasets.df
-    df.replace({np.nan:None}, inplace=True)  # replace NaN with None, otherwise we cannot convert it correctly
-    datasets_dict = df.to_dict(orient='records')   # is too stupid to handle NaN
-
-    return {
-        "datasets_dict": datasets_dict,
-        "totalCount": totalCount
-    }
-
-
-class SampleHandler(IPythonHandler):
-    """Handle the requests for /openbis/sample/connection/permId"""
-
-    def get(self, **params):
-        """Handle a request to /openbis/sample/connection_name/permId
-        download the dataset list and return a message
-        """
-        try:
-            conn = openbis_connections[params['connection_name']]
-        except KeyError:
-            self.set_status(500)
-            self.write({
-                "reason" : 'connection {} was not found'.format(
-                    params['connection_name']
-                )
-            })
-            return
-
-        if not conn.is_session_active():
-            try:
-                conn.login()
-            except Exception as exc:
-                self.set_status(500)
-                self.write({
-                    "reason" : 'connection to {} could not be established: {}'.format(conn.name, exc)
-                })
-                return
-
-        entity = get_entity_for_identifier(conn, params['identifier'])
-        if entity is None:
-            self.set_status(404)
-            self.write({
-                "reason" : 'No such Sample or Experiment: {}'.format(params['identifier'])
-            })
-            return None
-
-        querystring = parse_qs(self.request.query)
-        start_with = querystring.get('start_with', ['0'])[0]
-        count      = querystring.get('count', ['10'])[0]
-
-        datasets = get_datasets(entity, start_with=start_with, count=count)
-        if datasets is not None:
-            self.set_status(200)
-            self.write({
-                "dataSets"    : datasets.get('datasets_dict'),
-                "entity_attrs": entity.attrs.all(),
-                "entity_props": entity.props.all(),
-                "start_with"  : start_with,
-                "count"       : count,
-                "totalCount"  : datasets.get('totalCount'),
-                "cwd"         : os.getcwd()
-            })
-
diff --git a/jupyter-openbis-extension/server.py b/jupyter-openbis-extension/server.py
deleted file mode 100644
index 6d95bd59f0cf163c1bc92f74ede90df13e633634..0000000000000000000000000000000000000000
--- a/jupyter-openbis-extension/server.py
+++ /dev/null
@@ -1,135 +0,0 @@
-from notebook.utils import url_path_join
-import os
-import yaml
-
-from .connection import OpenBISConnections, OpenBISConnectionHandler, register_connection
-from .dataset import DataSetTypesHandler, DataSetDownloadHandler, DataSetUploadHandler
-from .sample import SampleHandler
-
-
-def _jupyter_server_extension_paths():
-    return [{'module': 'jupyter-openbis-extension.server'}]
-
-def _load_configuration(paths, filename='openbis-connections.yaml'):
-
-    if paths is None:
-        paths = []
-        home = os.path.expanduser("~")
-        paths.append(os.path.join(home, '.jupyter'))
-
-    # look in all config file paths of jupyter
-    # for openbis connection files and load them
-    connections = []
-    for path in paths:
-        abs_filename = os.path.join(path, filename)
-        if os.path.isfile(abs_filename):
-            with open(abs_filename, 'r') as stream:
-                try:
-                    config = yaml.safe_load(stream)
-                    for connection in config['connections']:
-                        connections.append(connection)
-                except yaml.YAMLexception as e:
-                    print(e)
-                    return None
-
-    return connections
-        
-
-def load_jupyter_server_extension(nb_server_app):
-    """Call when the extension is loaded.
-    :param nb_server_app: Handle to the Notebook webserver instance.
-    """
-
-    # load the configuration file
-    # and register the openBIS connections.
-    # If username and password is available, try to connect to the server
-    connections = _load_configuration(
-        paths    = nb_server_app.config_file_paths,
-        filename = 'openbis-connections.yaml'
-    )
-
-
-    for connection_info in connections:
-        conn = register_connection(connection_info)
-        print("Registered: {}".format(conn.url))
-        if conn.username and conn.password:
-            try:
-                conn.login()
-                print("Successfully connected to: {}".format(conn.url))
-            except ValueError:
-                print("Incorrect username or password for: {}".format(conn.url))
-            except Exception:
-                print("Cannot establish connection to: {}".format(conn.url))
-
-    # Add URL handlers to our web_app
-    # see Tornado documentation: https://www.tornadoweb.org
-    web_app = nb_server_app.web_app
-    host_pattern = '.*$'
-    base_url = web_app.settings['base_url']
-
-
-    # DataSet download
-    web_app.add_handlers(
-        host_pattern, 
-        [(url_path_join(
-            base_url,
-            '/openbis/dataset/(?P<connection_name>.*)?/(?P<permId>.*)?/(?P<downloadPath>.*)'), 
-            DataSetDownloadHandler
-        )]
-    )
-
-    # DataSet upload
-    web_app.add_handlers( host_pattern, [(
-            url_path_join( base_url, '/openbis/dataset/(?P<connection_name>.*)' ),
-            DataSetUploadHandler
-        )]
-    )
-
-    # DataSet-Types
-    web_app.add_handlers(
-        host_pattern,
-        [(
-            url_path_join( base_url, '/openbis/datasetTypes/(?P<connection_name>.*)'
-            ),
-            DataSetTypesHandler
-        )]
-    )
-
-    # DataSets for Sample identifier/permId
-    web_app.add_handlers(
-        host_pattern, 
-        [(  
-            url_path_join(
-                base_url,
-                '/openbis/sample/(?P<connection_name>.*)?/(?P<identifier>.*)'
-            ), 
-            SampleHandler
-        )]
-    )
-
-    # OpenBIS connections
-    web_app.add_handlers(
-        host_pattern,
-        [(
-            url_path_join(
-                base_url,
-                '/openbis/conns'
-            ),
-            OpenBISConnections
-        )]
-    )
-
-    # Modify / reconnect to a connection
-    web_app.add_handlers(
-        host_pattern, 
-        [(
-            url_path_join(
-                    base_url,
-                    '/openbis/conn/(?P<connection_name>.*)'
-            ),
-            OpenBISConnectionHandler
-        )]
-    )
-
-
-
diff --git a/jupyter-openbis-extension/static/state.js b/jupyter-openbis-extension/static/state.js
index bcbe83fe121570577138ac2a503ac24d91bd5951..18868d2309bf9b0dd738587e3e3f534b4d8852bb 100644
--- a/jupyter-openbis-extension/static/state.js
+++ b/jupyter-openbis-extension/static/state.js
@@ -1,5 +1,6 @@
 define([],
     function () {
+        // noinspection JSAnnotator
         return {
             // connection dialog
             connection: {
@@ -26,7 +27,15 @@ define([],
             entity: null,
 
             // openBIS v3 connection
-            openbisService : null
+            openbisService : null,
+
+            //runtime environment (e.g. Python version) and requirements (package list)
+
+            requirements_list : null,
+            requirements_filename : null,
+            runtime_filename : null,
+            runtime : null
+
         }
     }
 )
\ No newline at end of file
diff --git a/jupyter-openbis-extension/static/uploadDialog.js b/jupyter-openbis-extension/static/uploadDialog.js
index 16e86669d0fbbee5cee2bafd112a66ea0b135591..096d0514e6fbc7b6992801f8b074a0b7e582c49e 100644
--- a/jupyter-openbis-extension/static/uploadDialog.js
+++ b/jupyter-openbis-extension/static/uploadDialog.js
@@ -270,20 +270,25 @@ define([
                 filenameCell.style.width = "100%"
                 filenameCell.style.cursor = "pointer"
 
+                var checkbox = document.createElement("INPUT")
+                checkbox.type = "checkbox"
+                checkbox.value = file.path
+                checkbox.checked = state.selectedFiles.includes(file.path) ? true: false
+                checkbox.onclick = registerFile
+                checkboxes.push(checkbox)
+                checkboxCell.appendChild(checkbox)
+
                 if (file.type === "directory") {
                     iconCell.className = "item_icon folder_icon icon-fixed-width"
                     filenameCell.onclick = function () {
                         get_file_list(env, container, file.path)
                     }
+                    iconCell.onclick = function () {
+                        checkbox.checked = !checkbox.checked
+                        registerFile.call(checkbox)
+                    }
                 }
                 else {
-                    var checkbox = document.createElement("INPUT")
-                    checkbox.type = "checkbox"
-                    checkbox.value = file.path
-                    checkbox.checked = state.selectedFiles.includes(file.path) ? true: false
-                    checkbox.onclick = registerFile
-                    checkboxes.push(checkbox)
-                    checkboxCell.appendChild(checkbox)
 
                     if (file.type === "notebook") {
                         iconCell.className = "item_icon notebook_icon icon-fixed-width"
@@ -409,7 +414,7 @@ define([
                     if (! files.includes(env.notebook.notebook_path)) {
                         files.push(env.notebook.notebook_path)
                     }
-                    console.log(files)
+                    //console.log(files)
 
                     var props = {}
                     for (input of $('#upload-input-fields').find('input')) {
@@ -494,6 +499,89 @@ define([
                     keyboard_manager: env.notebook.keyboard_manager
                 }
 
+                var notebook = IPython.notebook
+
+                var code_for_requirements = ""
+                var requirements_filename = ""
+                var code_for_runtime = ""
+                var runtime_filename = ""
+                if (notebook.metadata.kernelspec.language == "python") {
+                    code_for_requirements = "import pkg_resources; print(\"\\n\".join([\"{}=={}\".format(i.key, i.version) for i in pkg_resources.working_set]))"
+                    code_for_runtime = "import sys; print('python-' + str(sys.version_info[0]) + '.' + str(sys.version_info[1]))"
+                    requirements_filename = "requirements.txt"
+                    runtime_filename = "runtime.txt"
+                }
+                else if (notebook.metadata.kernelspec.language == "R") {
+                    code_for_requirements = ""
+                    requirements_filename = "runtime.txt"
+                }
+                function save_requirements(data) {
+                    notebook.metadata.language_info.requirements_list = data
+                    notebook.metadata.language_info.requirements_filename = requirements_filename
+                    state.requirements_list = data
+                    state.requirements_filename = requirements_filename
+                }
+                function save_runtime(data) {
+                    notebook.metadata.language_info.runtime = data
+                    notebook.metadata.language_info.runtime_filename = runtime_filename
+                    state.runtime = data
+                    state.runtime_filename = runtime_filename
+                }
+                var req_callback = {
+                    iopub: {
+                        output: (data) => save_requirements(data.content.text.trim())
+                    }
+                };
+                var rt_callback = {
+                    iopub: {
+                        output: (data) => save_runtime(data.content.text.trim())
+                    }
+                };
+
+                var kernel = IPython.notebook.kernel
+                kernel.execute(code_for_requirements, req_callback)
+                kernel.execute(code_for_runtime, rt_callback)
+
+
+                function send_runtime_requirements(state) {
+                    var endpoint = env.notebook.base_url + 'requirements'
+                    var notebook_path = IPython.notebook.notebook_path
+
+                    body = {
+                        "notebook_path": notebook_path,
+                        "requirements_list": state.requirements_list,
+                        "requirements_filename": state.requirements_filename,
+                        "runtime": state.runtime,
+                        "runtime_filename": state.runtime_filename
+                    }
+
+                    var xsrf_token = common.getCookie('_xsrf')
+                    fetch(endpoint, {
+                        method: "POST",
+                        headers: {
+                            "Content-Type": "application/json",
+                            "X-XSRFToken": xsrf_token,
+                            "credentials": "same-origin",
+                        },
+                        body: JSON.stringify(body)
+                    })
+                        .then(function (response) {
+                            if (response.ok) {
+                                //alert(response.status)
+                            }
+                            else {
+                                //alert(response.status)
+                            }
+
+                        })
+                        .catch(
+                            error => console.error(
+                                "Error while attempting to write requirement files: ", error
+                            )
+                        )
+                }
+                send_runtime_requirements(state)
+
                 if (env.notebook.dirty === true) {
                     env.notebook.save_notebook()
                         .then(function () {
diff --git a/setup.py b/setup.py
index dc56ee882d12e85791ccce9874f138aa3b2ee854..e238a78aaaaa8e992db7404cba9c84d7176b7ce9 100644
--- a/setup.py
+++ b/setup.py
@@ -11,7 +11,7 @@ with open("README.md", "r", encoding="utf-8") as fh:
 
 setup(
     name='jupyter-openbis-extension',
-    version= '0.4.0',
+    version= '0.5.0',
     author='Swen Vermeul |  ID SIS | ETH Zürich',
     author_email='swen@ethz.ch',
     description='Extension for Jupyter notebooks to connect to openBIS and download/upload datasets, inluding the notebook itself',
@@ -23,7 +23,7 @@ setup(
     install_requires=[
         'jupyter-nbextensions-configurator',
         'jupyter',
-        'pybis>=1.9.5',
+        'jupyter-openbis-server',
         'numpy',
         'tornado==5.1.1',
     ],
@@ -113,10 +113,6 @@ setup(
         ("etc/jupyter/nbconfig/notebook.d", [
             "jupyter-config/nbconfig/notebook.d/jupyter_openbis_extension.json"
         ]),
-        # like `jupyter serverextension enable --sys-prefix`
-        ("etc/jupyter/jupyter_notebook_config.d", [
-            "jupyter-config/jupyter_notebook_config.d/jupyter_openbis_extension.json"
-        ])
     ],
     zip_safe=False,
 )