How to use Galaxy?
In the left panel of this page you will find the list of available tools, you can see this page at any point by clicking ‘Analyze Data’ in the top menu.
The history of your current data analysis is shown in the right panel. A good practice when running multiple analysis in parallel is to create several histories and give each one a unique name. You can click on ‘View all histories’ button (upper right corner in histories panel) to get an overview and switch between the current histories.
Every analysis starts with getting the input data into your current history (right panel). To do this you can upload your own input files or use shared datasets. To upload files from your computer or instruct Galaxy to download files from the web you have to use of the upload tool: Get data (tool panel on the left) → upload file. Please DO NOT UPLOAD LARGE FILES (~GB), but contact the administrator to create a central data repository! To use available shared data you have to click on Shared data (top menu) → Data Libraries. You can then browse the available libraries and select the file/s you want to use. By clicking ‘to History’ button and choosing the desired History name you will import these files and make them available to use as input for future analysis.
To run a job select the tool from tool panel on the left, then the corresponding interface will be loaded and you will be able to select corresponding input data and (re)define parameters. Please read carefully the labels and help text next to the input fields. After clicking Execute you will be able to see entries for each in your history. The color of a dataset designates the current status of the underlying job
|Gray||The job is being evaluated to run (new dataset) or is queued. Allow this to complete.|
|Yellow||The job is executing.|
|Red||The job has failed.|
|Blue||The job is paused. This indicates either a problem with an input (a previous step in the workflow may have failed) or that you have exceeded disk quota set by the administrator of the Galaxy instance you are working on.|
You can automate your analysis pipeline by using workflows composed of several tools linked by their input/output data. In the Workflows section (upper menu) you can see a list of current workflows and also create your own ones. It is also possible (and highly recommended) to use the shared, and widely tested, workflows available under Shared Data (top menu) → workflows. To use one of these you first need to import it to your workflows list by clicking on the name and selecting Import. To execute any workflow listed under your workflows first click on it and select run, then choose the input data as with any tool.
Apr 3, 2021
Training Infrastructure Feedback from Simon Gladman
Mar 28, 2021
Training Infrastructure Feedback from Dr. Fotis Psomopoulos
Mar 25, 2021
Galaxy-ELIXIR webinars series: Advanced Features
Mar 25, 2021
Training Infrastructure Feedback from Dr. Ricardo Gonzalo Sanz
Mar 25, 2021
How to import a workflow from the Workflow Hub and run it on remote COVID-19 data
Mar 22, 2021
Training Infrastructure Feedback from Martin Čech
Oct 19, 2021
Analyse avancée de séquences
Oct 11, 2021
Workflow4Experimenters (W4E) international course 2021
Sep 14, 2021
Analyse de données de métabarcoding
Jul 25, 2021 - Jul 31, 2021
ISMB/ECCB 2021 & BOSC 2021
Jun 28, 2021 - Jul 10, 2021
Galaxy Community Conference 2021
Jun 7, 2021