skip to main content


Title: Engineering heterologous enzyme secretion in Yarrowia lipolytica
Abstract Background

Eukaryotic cells are often preferred for the production of complex enzymes and biopharmaceuticals due to their ability to form post-translational modifications and inherent quality control system within the endoplasmic reticulum (ER). A non-conventional yeast species,Yarrowia lipolytica, has attracted attention due to its high protein secretion capacity and advanced secretory pathway. Common means of improving protein secretion inY. lipolyticainclude codon optimization, increased gene copy number, inducible expression, and secretory tag engineering. In this study, we develop effective strategies to enhance protein secretion using the model heterologous enzyme T4 lysozyme.

Results

By engineering the commonly used native lip2prepro secretion signal, we have successfully improved secreted T4 lysozyme titer by 17-fold. Similar improvements were measured for other heterologous proteins, including hrGFP and$$\alpha$$α-amylase. In addition to secretion tag engineering, we engineered the secretory pathway by expanding the ER and co-expressing heterologous enzymes in the secretion tag processing pathway, resulting in combined 50-fold improvement in T4 lysozyme secretion.

Conclusions

Overall, our combined strategies not only proved effective in improving the protein production inYarrowia lipolytica, but also hint the possible existence of a different mechanism of secretion regulation in ER and Golgi body in this non-conventional yeast.

 
more » « less
NSF-PAR ID:
10368503
Author(s) / Creator(s):
;
Publisher / Repository:
Springer Science + Business Media
Date Published:
Journal Name:
Microbial Cell Factories
Volume:
21
Issue:
1
ISSN:
1475-2859
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract Background

    Rapid improvements in inexpensive, low-power, movement and environmental sensors have sparked a revolution in animal behavior research by enabling the creation of data loggers (henceforth, tags) that can capture fine-grained behavioral data over many months. Nevertheless, development of tags that are suitable for use with small species, for example, birds under 25 g, remains challenging because of the extreme mass (under 1$$\textrm{g}$$g) and power (average current under 1$$\upmu$$μA) constraints. These constraints dictate that a tag should carry exactly the sensors required for a given experiment and the data collection protocol should be specialized to the experiment. Furthermore, it can be extremely challenging to design hardware and software to achieve the energy efficiency required for long tag life.

    Results

    We present an activity monitor, BitTag, that can continuously collect activity data for 4–12 months at 0.5–0.8$$\textrm{g}$$g, depending upon battery choice, and which has been used to collect more than 500,000 h of data in a variety of experiments. The BitTag architecture provides a general platform to support the development and deployment of custom sub-$$\textrm{g}$$gtags. This platform consists of a flexible tag architecture, software for both tags and host computers, and hardware to provide the host/tag interface necessary for preparing tags for “flight” and for accessing tag data “post-flight”. We demonstrate how the BitTag platform can be extended to quickly develop novel tags with other sensors while satisfying the 1g/1$$\upmu$$μA mass and power requirements through the design of a novel barometric pressure sensing tag that can collect pressure and temperature data every 60$$\textrm{s}$$sfor a year with mass under 0.6$$\textrm{g}$$g.

     
    more » « less
  2. Abstract Background

    Protein–protein interaction (PPI) is vital for life processes, disease treatment, and drug discovery. The computational prediction of PPI is relatively inexpensive and efficient when compared to traditional wet-lab experiments. Given a new protein, one may wish to find whether the protein has any PPI relationship with other existing proteins. Current computational PPI prediction methods usually compare the new protein to existing proteins one by one in a pairwise manner. This is time consuming.

    Results

    In this work, we propose a more efficient model, called deep hash learning protein-and-protein interaction (DHL-PPI), to predict all-against-all PPI relationships in a database of proteins. First, DHL-PPI encodes a protein sequence into a binary hash code based on deep features extracted from the protein sequences using deep learning techniques. This encoding scheme enables us to turn the PPI discrimination problem into a much simpler searching problem. The binary hash code for a protein sequence can be regarded as a number. Thus, in the pre-screening stage of DHL-PPI, the string matching problem of comparing a protein sequence against a database withMproteins can be transformed into a much more simpler problem: to find a number inside a sorted array of lengthM. This pre-screening process narrows down the search to a much smaller set of candidate proteins for further confirmation. As a final step, DHL-PPI uses the Hamming distance to verify the final PPI relationship.

    Conclusions

    The experimental results confirmed that DHL-PPI is feasible and effective. Using a dataset with strictly negative PPI examples of four species, DHL-PPI is shown to be superior or competitive when compared to the other state-of-the-art methods in terms of precision, recall or F1 score. Furthermore, in the prediction stage, the proposed DHL-PPI reduced the time complexity from$$O(M^2)$$O(M2)to$$O(M\log M)$$O(MlogM)for performing an all-against-all PPI prediction for a database withMproteins. With the proposed approach, a protein database can be preprocessed and stored for later search using the proposed encoding scheme. This can provide a more efficient way to cope with the rapidly increasing volume of protein datasets.

     
    more » « less
  3. Abstract

    A wide range of dark matter candidates have been proposed and are actively being searched for in a large number of experiments, both at high (TeV) and low (sub meV) energies. One dark matter candidate, a deeply bounduuddsssexaquark,$$S$$S, with mass$$\sim 2$$2GeV (having the same quark content as the hypothesized H-dibaryon, but long lived) is particularly difficult to explore experimentally. In this paper, we propose a scheme in which such a state could be produced at rest through the formation of$$\bar{p}$$p¯$$^3$$3He antiprotonic atoms and their annihilation into$$S$$S+$$K^+K^+\pi ^-$$K+K+π-, identified both through the unique tag of a$$S=+2, Q=+1$$S=+2,Q=+1final state, as well as through full kinematic reconstruction of the final state recoiling against it.

     
    more » « less
  4.  
    more » « less
  5. Abstract Background

    Many bioinformatics applications involve bucketing a set of sequences where each sequence is allowed to be assigned into multiple buckets. To achieve both high sensitivity and precision, bucketing methods are desired to assign similar sequences into the same bucket while assigning dissimilar sequences into distinct buckets. Existingk-mer-based bucketing methods have been efficient in processing sequencing data with low error rates, but encounter much reduced sensitivity on data with high error rates. Locality-sensitive hashing (LSH) schemes are able to mitigate this issue through tolerating the edits in similar sequences, but state-of-the-art methods still have large gaps.

    Results

    In this paper, we generalize the LSH function by allowing it to hash one sequence into multiple buckets. Formally, a bucketing function, which maps a sequence (of fixed length) into a subset of buckets, is defined to be$$(d_1, d_2)$$(d1,d2)-sensitive if any two sequences within an edit distance of$$d_1$$d1are mapped into at least one shared bucket, and any two sequences with distance at least$$d_2$$d2are mapped into disjoint subsets of buckets. We construct locality-sensitive bucketing (LSB) functions with a variety of values of$$(d_1,d_2)$$(d1,d2)and analyze their efficiency with respect to the total number of buckets needed as well as the number of buckets that a specific sequence is mapped to. We also prove lower bounds of these two parameters in different settings and show that some of our constructed LSB functions are optimal.

    Conclusion

    These results lay the theoretical foundations for their practical use in analyzing sequences with high error rates while also providing insights for the hardness of designing ungapped LSH functions.

     
    more » « less