prompt
stringlengths
1.06k
15.6k
ori_text
stringlengths
1.02k
15.6k
reference_list
stringlengths
65
2.27k
__internal_uuid__
stringlengths
5
5
Primary_Domain
stringclasses
4 values
Secondary_Domain
stringclasses
17 values
prompt_id
stringlengths
1
3
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: The preceding sections have examined various theories on concepts and representations in philosophy, cognitive science, and machine learning, classifying them into four categories: abstractionist, similarity, functional, and invariance approaches. This section compares these views from a meta-perspective, exploring their connections and deriving implications for further studies. By identifying the intersections and differences among these approaches, this discussion section aims to provide a clearer framework for interdisciplinary studies of concepts and representations. The first axis for comparison is the role of concepts. Arguably, concepts play various roles in classification, learning, communication, problem solving, and so on; but these roles can be further understood through the lens of two major functionalities: descriptive and inferential. In the descriptive use, concepts serve to characterize and summarize data in various formats. Among the approaches discussed so far, the abstractionist and similarity approaches are particularly motivated by these descriptive tasks. The major goal of abstractionists is to classify objects in the hierarchical manner, by constructing a conceptual lattice from a data table or “context” that lists items and their observed properties (Section 2). Given a similar (but possibly continuous) dataset, the similarity approach visualizes mutual relationships between items in a two or more dimensional space based on their attributes and delineates concepts as clustered points that are close to each other (Section 3). These outcomes—lattices or plots—reveal the inherent structure of data not visible in the original format. Concepts also play a central role in inferential tasks, e.g., for the purpose of predicting future events or possible behaviors of objects. By identifying an object in front of me as a dog, I can anticipate what it will and will not do: it may bark, chase after balls, and sniff around, but probably not climb trees or purr like a cat. Though very simplistic, it is nonetheless a bona fide act of prediction: based on some observable cues like floppy ears, I derive its behavioral characteristics that I have not yet seen, at least with respect to this particular dog in front of me. Such inferential role of concepts is the primary focus of the functional perspective (Section 4). Inferences from one set of attributes to another exploit the information about internal constraints on the possible combinations of attributes. This information, encoded in the functional form, constitutes our understanding of concepts. Alternatively, the invariance approach captures inference from a different perspective, by identifying a concept as a set of trans- formation rules that govern possible changes in the object (Section 5). Such equivariant group actions highlight the inferential role of concepts that allows us to simulate how objects or their perceptions transform in accordance with changes in the environment or the perceiving subject. The above discussion should not imply that models are inherently connected to or designed for particular roles. Indeed, there is no contradiction in taking a conceptual lattice as representing internal constraints of concepts or using the similarity space for the purpose of prediction, as is quite common in natural language processing. The descriptive/inferential contrast rather concerns the modeling purpose: namely whether one is interested in data themselves, or in some underlying structure of “uniformity of nature” from which data are sampled. In the former context, mathematical models are taken as a sort of descriptive statistics that serve for the economy of thought; while in the latter, they are interpreted as representing a generative mechanism that lies beyond any particular data (Otsuka, 2022). These two desiderata are often in conflict: the more one tries to adjusts one’s concepts as faithful as possible to given data, the more likely they are to overfit the data, resulting in a loss of predictive ability (Forster and Sober, 1994). A good descriptive model needs not be a good inferential model, and vice versa. Therefore, when evaluating a particular model, it is essential to clearly specify the criteria by which it is being evaluated. Our second point of consideration concerns the relationship between concepts and their defining features. In the philosophical and cognitive science literature, concepts have been defined or characterized in terms of familiar and explicit features we use in everyday life, such as color, size, shape, and so on. Accordingly, conceptual hierarchies or similarity maps have been build and evaluated on the basis of a pre-specified set of mostly ostensible features. This represents a “feature-first” approach, where features act as the raw material for thoughts, with concepts being formed by combining or mixing these features. In contrast, features in representation learning are not fixed beforehand, but rather are extracted from data as axes or dimensions that constitute the latent space. In addition, the meaning of these extracted features is not given a priori, but must be determined by unpacking the internal intricate structure of the trained neural network through detailed analysis. This can be seen as a “representation-first” approach, considering representations to be epistemologically prior to features. These two approach faces different challenges. The main challenge of the feature-first approach lies in its empirical adequacy. A common criticism against abstractionism or the classic view of concepts points out that it is simply im- possible to define concepts, such as game or human, through combinations of existing features or traits (Wittgenstein, 1953; Boyd, 1991). Moreover, it is not clear which features should be considered to define concepts. Outcomes of formal concept analysis heavily depend on the list of attributes used to characterize data (Section 2). And Feigenbaum’s bottleneck highlights the difficulty of determining the relevant attributes for the problem at hand. A similar issue arises for prototype and exemplar theorists in determining the appropriate at- tributes/dimensions and their relative importance in constructing the similarity space, which significantly affects similarity judgments (Murphy, 2004). With its successful applications to a range of empirical problems, the representation- first approach seems to be overcoming all the above issues: deep learning models are able to learn robust representations (with some reservations discussed shortly) that automatically extract relevant features from data, capture complex patterns, and generalize well to new situations. The problem, however, is that these features generally resist intuitive interpretation. One thus needs to read off meanings from trained models, but to do so requires a clear understanding of what one is looking for—that is, what are meaningful features? The idea of disentangled representation discussed in Section 5 is one attempt to explicate what we consider meaningful features of a representation in terms of independent group actions. Understanding features is crucial not just for interpretability but also to enhance model performance, ensure robustness, and improve fair- ness in machine learning applications (Lipton, 2016). It is pointed out that the well-known phenomenon of adversarial attack, where deep learning models mis- classify objects due to the addition of small, often imperceptible perturbations to the input data, is a consequence of the model’s using complex, high-dimensional features that are not necessarily aligned with human-perceptible features (Ilyas et al., 2019). Also, understanding features allows data scientists to identify and mitigate biases in machine learning models by examining how different features influence predictions (Ribeiro et al., 2016). These efforts can be understood as attempts to identify the concepts (representations) used by machines and analyze them into components (features) amenable to human understanding. The feature-first and representation-first approaches can thus be understood as akin to digging a tunnel from opposite sides. The former starts with a set of explicit features and builds complex concepts in a bottom-up way, while the latter aims to break down given representations into understandable pieces in a top-down fashion. The important challenge in contemporary concept research is to make these approaches meet halfway. The final, but not least, point concerns the relationship among the four approaches discussed in this paper. In his influential book, Edouard Machery (2009) argued for the disunity of concepts, challenging the traditional view that concepts are a unified phenomenon within cognitive science. His view is that what are labeled “concepts” actually involve heterogeneous mental kinds with different functionalities, purposes, and empirical bases. Be that as it may, mathematical considerations naturally suggest the logical relationships between different conceptual models. Indeed, we have already seen some of such attempts in the present paper. The group-theoretic analysis of disentangled representations can be thought as an attempt to integrate the theoretic aspect of concepts (encoded by group operations) and their similarity-based aspect (represented by a manifold). In Section 3, we have seen some recent works in natural language processing that aim to encode the hierarchical structure of concepts into the word vector space by using non-Euclidean (hyperbolic) spaces (Nickel and Kiela, 2017) or representing words by boxes instead of vectors (Vilnis et al., 2018). If successful, this line of research will reconcile the abstractionist and similarity approaches, which have been considered rivals in both philosophy and cognitive science literature. Underlying these studies is the overarching theme of the relationship be- tween geometry and algebra. Lattices and groups are algebraic in nature, while metric spaces and manifolds have a clear geometric character. Hence the four approaches discussed in this paper can each be seen as shedding light on the geometric or algebraic aspects of concepts. Kant (1999) was the first to make a clear distinction between, and propose a unification of, these two aspects of human cognition—namely sensibility equipped with a geometric form, and understanding that follows logical and algebraic principles. Over two centuries after Kant, the contemporary machine learning research is trying to integrate the both components to understand and improve the performance of neural networks, with the aid of much more advanced mathematical machinery than those available to Kant, including non-Euclidean geometry, topology, and group or gauge theory (Sanborn et al., 2024). Just as Kant was inspired by Newtonian physics in his time, these developments in machine learning will provide new insights into the philosophical understanding of concepts.
The preceding sections have examined various theories on concepts and representations in philosophy, cognitive science, and machine learning, classifying them into four categories: abstractionist, similarity, functional, and invariance approaches. This section compares these views from a meta-perspective, exploring their connections and deriving implications for further studies. By identifying the intersections and differences among these approaches, this discussion section aims to provide a clearer framework for interdisciplinary studies of concepts and representations. The first axis for comparison is the role of concepts. Arguably, concepts play various roles in classification, learning, communication, problem solving, and so on; but these roles can be further understood through the lens of two major functionalities: descriptive and inferential. In the descriptive use, concepts serve to characterize and summarize data in various formats. Among the approaches discussed so far, the abstractionist and similarity approaches are particularly motivated by these descriptive tasks. The major goal of abstractionists is to classify objects in the hierarchical manner, by constructing a conceptual lattice from a data table or “context” that lists items and their observed properties (Section 2). Given a similar (but possibly continuous) dataset, the similarity approach visualizes mutual relationships between items in a two or more dimensional space based on their attributes and delineates concepts as clustered points that are close to each other (Section 3). These outcomes—lattices or plots—reveal the inherent structure of data not visible in the original format. Concepts also play a central role in inferential tasks, e.g., for the purpose of predicting future events or possible behaviors of objects. By identifying an object in front of me as a dog, I can anticipate what it will and will not do: it may bark, chase after balls, and sniff around, but probably not climb trees or purr like a cat. Though very simplistic, it is nonetheless a bona fide act of prediction: based on some observable cues like floppy ears, I derive its behavioral characteristics that I have not yet seen, at least with respect to this particular dog in front of me. Such inferential role of concepts is the primary focus of the functional perspective (Section 4). Inferences from one set of attributes to another exploit the information about internal constraints on the possible combinations of attributes. This information, encoded in the functional form, constitutes our understanding of concepts. Alternatively, the invariance approach captures inference from a different perspective, by identifying a concept as a set of trans- formation rules that govern possible changes in the object (Section 5). Such equivariant group actions highlight the inferential role of concepts that allows us to simulate how objects or their perceptions transform in accordance with changes in the environment or the perceiving subject. The above discussion should not imply that models are inherently connected to or designed for particular roles. Indeed, there is no contradiction in taking a conceptual lattice as representing internal constraints of concepts or using the similarity space for the purpose of prediction, as is quite common in natural language processing. The descriptive/inferential contrast rather concerns the modeling purpose: namely whether one is interested in data themselves, or in some underlying structure of “uniformity of nature” from which data are sampled. In the former context, mathematical models are taken as a sort of descriptive statistics that serve for the economy of thought; while in the latter, they are interpreted as representing a generative mechanism that lies beyond any particular data (Otsuka, 2022). These two desiderata are often in conflict: the more one tries to adjusts one’s concepts as faithful as possible to given data, the more likely they are to overfit the data, resulting in a loss of predictive ability (Forster and Sober, 1994). A good descriptive model needs not be a good inferential model, and vice versa. Therefore, when evaluating a particular model, it is essential to clearly specify the criteria by which it is being evaluated. Our second point of consideration concerns the relationship between concepts and their defining features. In the philosophical and cognitive science literature, concepts have been defined or characterized in terms of familiar and explicit features we use in everyday life, such as color, size, shape, and so on. Accordingly, conceptual hierarchies or similarity maps have been build and evaluated on the basis of a pre-specified set of mostly ostensible features. This represents a “feature-first” approach, where features act as the raw material for thoughts, with concepts being formed by combining or mixing these features. In contrast, features in representation learning are not fixed beforehand, but rather are extracted from data as axes or dimensions that constitute the latent space. In addition, the meaning of these extracted features is not given a priori, but must be determined by unpacking the internal intricate structure of the trained neural network through detailed analysis. This can be seen as a “representation-first” approach, considering representations to be epistemologically prior to features. These two approach faces different challenges. The main challenge of the feature-first approach lies in its empirical adequacy. A common criticism against abstractionism or the classic view of concepts points out that it is simply im- possible to define concepts, such as game or human, through combinations of existing features or traits (Wittgenstein, 1953; Boyd, 1991). Moreover, it is not clear which features should be considered to define concepts. Outcomes of formal concept analysis heavily depend on the list of attributes used to characterize data (Section 2). And Feigenbaum’s bottleneck highlights the difficulty of determining the relevant attributes for the problem at hand. A similar issue arises for prototype and exemplar theorists in determining the appropriate at- tributes/dimensions and their relative importance in constructing the similarity space, which significantly affects similarity judgments (Murphy, 2004). With its successful applications to a range of empirical problems, the representation- first approach seems to be overcoming all the above issues: deep learning models are able to learn robust representations (with some reservations discussed shortly) that automatically extract relevant features from data, capture complex patterns, and generalize well to new situations. The problem, however, is that these features generally resist intuitive interpretation. One thus needs to read off meanings from trained models, but to do so requires a clear understanding of what one is looking for—that is, what are meaningful features? The idea of disentangled representation discussed in Section 5 is one attempt to explicate what we consider meaningful features of a representation in terms of independent group actions. Understanding features is crucial not just for interpretability but also to enhance model performance, ensure robustness, and improve fair- ness in machine learning applications (Lipton, 2016). It is pointed out that the well-known phenomenon of adversarial attack, where deep learning models mis- classify objects due to the addition of small, often imperceptible perturbations to the input data, is a consequence of the model’s using complex, high-dimensional features that are not necessarily aligned with human-perceptible features (Ilyas et al., 2019). Also, understanding features allows data scientists to identify and mitigate biases in machine learning models by examining how different features influence predictions (Ribeiro et al., 2016). These efforts can be understood as attempts to identify the concepts (representations) used by machines and analyze them into components (features) amenable to human understanding. The feature-first and representation-first approaches can thus be understood as akin to digging a tunnel from opposite sides. The former starts with a set of explicit features and builds complex concepts in a bottom-up way, while the latter aims to break down given representations into understandable pieces in a top-down fashion. The important challenge in contemporary concept research is to make these approaches meet halfway. The final, but not least, point concerns the relationship among the four approaches discussed in this paper. In his influential book, Edouard Machery (2009) argued for the disunity of concepts, challenging the traditional view that concepts are a unified phenomenon within cognitive science. His view is that what are labeled “concepts” actually involve heterogeneous mental kinds with different functionalities, purposes, and empirical bases. Be that as it may, mathematical considerations naturally suggest the logical relationships between different conceptual models. Indeed, we have already seen some of such attempts in the present paper. The group-theoretic analysis of disentangled representations can be thought as an attempt to integrate the theoretic aspect of concepts (encoded by group operations) and their similarity-based aspect (represented by a manifold). In Section 3, we have seen some recent works in natural language processing that aim to encode the hierarchical structure of concepts into the word vector space by using non-Euclidean (hyperbolic) spaces (Nickel and Kiela, 2017) or representing words by boxes instead of vectors (Vilnis et al., 2018). If successful, this line of research will reconcile the abstractionist and similarity approaches, which have been considered rivals in both philosophy and cognitive science literature. Underlying these studies is the overarching theme of the relationship be- tween geometry and algebra. Lattices and groups are algebraic in nature, while metric spaces and manifolds have a clear geometric character. Hence the four approaches discussed in this paper can each be seen as shedding light on the geometric or algebraic aspects of concepts. Kant (1999) was the first to make a clear distinction between, and propose a unification of, these two aspects of human cognition—namely sensibility equipped with a geometric form, and understanding that follows logical and algebraic principles. Over two centuries after Kant, the contemporary machine learning research is trying to integrate the both components to understand and improve the performance of neural networks, with the aid of much more advanced mathematical machinery than those available to Kant, including non-Euclidean geometry, topology, and group or gauge theory (Sanborn et al., 2024). Just as Kant was inspired by Newtonian physics in his time, these developments in machine learning will provide new insights into the philosophical understanding of concepts.
考点1:“Sensibility”推荐译为“感性” 考点2:“unpacking”推荐译为“解析”,不可译为“拆解” 考点3:“The first axis for comparison is the role of concepts”中“axis”不可直译为“轴”,应为“维度”或“方面” 考点4:“read off meanings”推荐译为“解读含义”“阐释含义”,不可译为“读出含义”
0689a
学术论文
人文科学
185
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: 1. Why has the global capital market grown so rapidly in recent decades In recent decades, the global capital market has grown so rapidly because of the rise of privatizations mainly. With private capital flows rising from less than 5 percent of world GDP in 1975 to about 20 percent today, privatizations have significantly increased market liquidity. And also privatization takes a potential role global capital market development. A. The Rise of Capital Market-Based Finance Capital market-based finance has in fact been increasing in importance, both absolutely and relative to financial intermediary-based finance, in both developed and developing countries over the past decade. And also capital markets are in fact winning the present and seem likely to dominate the future of corporate finance in developed and developing countries alike. a. The Stable Role of Commercial Banking in Modern Economies Ordinary "relationship banking" appears to be (at best) holding its own as a source of corporate financing around the world, and is more likely in decline. The bits of banking that are growing rapidly are those parts that provide high value-added products (especially risk management tools) and provide large-scale syndicated credits to corporate borrowers. During the late-1980s and early-1990s, when Japan and Germany appeared to be outperforming major capital market-oriented countries such as Britain and the US, the academic literature often favored bank-based systems. Examples of this literature include Prowse (1992), Kester (1992), and Porter (1992), while the supporting arguments are summarized in Maher and Andersson (1999) and Tsuru (2000). More recently, however, the weight of opinion has swung strongly in favor of the idea that capital markets have decisive comparative advantages over banks and other financial intermediaries as optimal monitors and financiers of a nation's corporate life. This reassessment has been driven in part by the observation, discussed at length above, that capital markets have been prospering relative to banks for many years now. The repetitive nature--and massive costs--of banking crises in developing and developed countries alike has also convinced many observers that banks are inherently fragile institutions, whose role in corporate finance should be minimized as much and as quickly as possible (Economist (1997, 1999)). b. The Rapid Growth in Stock Market Capitalization and Trading Volume Since 1983 From 1983 to 2000, this was a period of very rapid growth in the capitalization of markets in every country except Japan. Total world market capitalization increased over ten-fold (to $ 35.0 trillion) between 1983 and 1999, and the total capitalization of the US market increased almost nine-fold (from $ 1.9 trillion to $ 16.6 trillion) over the same period. c. The Dramatic Growth in Securities Issuance Volume Since 1990 Another way of measuring the rise of capital markets is to examine whether their share of annual corporate financing activity has grown relative to that of other sources of funding. Security offerings by US issuers accounted for two-thirds of the global total throughout 1990-1999, that implies that non-US securities issues in creased from $ 191 billion in 1990 to $ 750 billion in 1998, and then to $ 1.19 trillion in 1999. The surge in non-US issuance volume in 1999 was largely due to the popularity of euro-denominated bond issues, which actually exceeded dollar-denominated bond issues for much of 1999. d. The Phenomenal Growth in Venture Capital Financing in the United States One highly specialized, but extremely important type of financing has also grown very rapidly over the past decade, and especially so since 1997. This is venture capital investment by US venture capital partnerships. The fund-raising patterns of these private equity investors are discussed in Gompers and Lerner (1998), and the competitive advantages of US venture capitalists versus those in other developed countries are described in Black and Gilson (1998). e. The Surge in Mergers and Acquisitions Worldwide The almost incredible increase in the total volume of merger and acquisition activity that has occurred since 1990. While takeovers have always played an important role in the United States, the rise in M&A (Merger and Acquisition) activity in Europe during the 1990s was even more dramatic. From less than $ 50 billion annually in the late-1980s, the total value of M&A involving a European target reached $ 592 billion in 1998, before more than doubling to $ 1.22 trillion in 1999--rivaling the US total. The global value of M&A activity in 1999 reached $ 3.4 trillion, an astounding 10% of world GDP. Next I will document that share issue privatizations have truly transformed share ownership patterns of investors in many different countries. B. Privatization's Impact on Stock and Bond Market Development We should be careful in inferring causation regarding privatization's impact on market growth, since a shift in ideology or some other exogenous political or economic change might have caused both the privatization and the overall boom. a. Total Proceeds Raised by Privatization Programs It is clear that national governments have been among the biggest winners from privatization programs, since these have dramatically increased government revenues, which is clearly one reason the policy has spread so rapidly. As mentioned above, Privatisation International [Gibbon (1998, 2000)] reports that the cumulative value of proceeds raised by privatizing governments exceeded $ 1 trillion sometime during the second half of 1999. As an added benefit, this revenue has come to governments without having to raise taxes or cut other public services. b. Privatization's Impact on International Investment Banking All international investment banks compete fiercely for share issue privatization mandates, for two principal reasons. First, because the offerings are so large and so visible--and are almost always designed to help promote the market's capacity to absorb subsequent stock offerings by private companies--these are very prestigious mandates. To date, the large US and British brokerage houses have had the most success in winning advisory and underwriting mandates, though all countries that launch large-scale SIP programs tend to favor local investment banks as "national champions" to handle the domestic share tranche. The second reason banks compete so fiercely for SIP mandates is because they can be extremely profitable. In spite of the fact--documented by Jones, et al (1999) and Ljungqvist, et al (2000)--that SIPs have significantly lower underwriting spreads than private sector offerings, their sheer size and lack of downside price risk make them very lucrative for underwriters. 2. Will this growth continue throughout the 2000s? As we indicated above, the global capital market has grown so rapidly in recent decades cause of the privatizations rise. Privatizations increased the market liquidity. Now we have already stepped into the 21st century. I believe that the growth will continue for the following reasons. First, most of the south-east Asia countries have recovered from the 1997 financial crisis. For these countries, they now have the capital to do businesses. And they get back on the fast growing track. Second, by the end of 2001, world's biggest developing country, China, has entered the WTO (World Trade Organization). This is real great news. As we all know, today's China takes a serious position in world's economy. Its innovation and opening policy make china keep achieving high GDP growth rate. This drives the global capital market keep growing. Summary and Conclusions This essay examines the impact of share issue privatizations (SIPs) on the growth of world capital markets (especially stock markets). I begin by documenting the increasing importance of capital markets, and the declining role of commercial banks, in corporate financial systems around the world. I then show that privatization programs-- particularly those involving public share offerings--have had a dramatic impact both on the development of non-US stock markets and on the participation of individual and institutional investors in those stock markets. This has told the reason of the fast growth of global capital market. And then I succinctly indicated the continuance of the rapid growth, the great future. The last but not the least is the recommendation. I can confidently assert that, if executed properly, a series of share issue privatizations can indeed promote the growth of global capital market, which will yield economic and political dividends for many years to come. That means there is a need to encourage the development of SIPs in order to gain growth of global capital market.
1. Why has the global capital market grown so rapidly in recent decades In recent decades, the global capital market has grown so rapidly because of the rise of privatizations mainly. With private capital flows rising from less than 5 percent of world GDP in 1975 to about 20 percent today, privatizations have significantly increased market liquidity. And also privatization takes a potential role global capital market development. A. The Rise of Capital Market-Based Finance Capital market-based finance has in fact been increasing in importance, both absolutely and relative to financial intermediary-based finance, in both developed and developing countries over the past decade. And also capital markets are in fact winning the present and seem likely to dominate the future of corporate finance in developed and developing countries alike. a. The Stable Role of Commercial Banking in Modern Economies Ordinary "relationship banking" appears to be (at best) holding its own as a source of corporate financing around the world, and is more likely in decline. The bits of banking that are growing rapidly are those parts that provide high value-added products (especially risk management tools) and provide large-scale syndicated credits to corporate borrowers. During the late-1980s and early-1990s, when Japan and Germany appeared to be outperforming major capital market-oriented countries such as Britain and the US, the academic literature often favored bank-based systems. Examples of this literature include Prowse (1992), Kester (1992), and Porter (1992), while the supporting arguments are summarized in Maher and Andersson (1999) and Tsuru (2000). More recently, however, the weight of opinion has swung strongly in favor of the idea that capital markets have decisive comparative advantages over banks and other financial intermediaries as optimal monitors and financiers of a nation's corporate life. This reassessment has been driven in part by the observation, discussed at length above, that capital markets have been prospering relative to banks for many years now. The repetitive nature--and massive costs--of banking crises in developing and developed countries alike has also convinced many observers that banks are inherently fragile institutions, whose role in corporate finance should be minimized as much and as quickly as possible (Economist (1997, 1999)). b. The Rapid Growth in Stock Market Capitalization and Trading Volume Since 1983 From 1983 to 2000, this was a period of very rapid growth in the capitalization of markets in every country except Japan. Total world market capitalization increased over ten-fold (to $ 35.0 trillion) between 1983 and 1999, and the total capitalization of the US market increased almost nine-fold (from $ 1.9 trillion to $ 16.6 trillion) over the same period. c. The Dramatic Growth in Securities Issuance Volume Since 1990 Another way of measuring the rise of capital markets is to examine whether their share of annual corporate financing activity has grown relative to that of other sources of funding. Security offerings by US issuers accounted for two-thirds of the global total throughout 1990-1999, that implies that non-US securities issues in creased from $ 191 billion in 1990 to $ 750 billion in 1998, and then to $ 1.19 trillion in 1999. The surge in non-US issuance volume in 1999 was largely due to the popularity of euro-denominated bond issues, which actually exceeded dollar-denominated bond issues for much of 1999. d. The Phenomenal Growth in Venture Capital Financing in the United States One highly specialized, but extremely important type of financing has also grown very rapidly over the past decade, and especially so since 1997. This is venture capital investment by US venture capital partnerships. The fund-raising patterns of these private equity investors are discussed in Gompers and Lerner (1998), and the competitive advantages of US venture capitalists versus those in other developed countries are described in Black and Gilson (1998). e. The Surge in Mergers and Acquisitions Worldwide The almost incredible increase in the total volume of merger and acquisition activity that has occurred since 1990. While takeovers have always played an important role in the United States, the rise in M&A (Merger and Acquisition) activity in Europe during the 1990s was even more dramatic. From less than $ 50 billion annually in the late-1980s, the total value of M&A involving a European target reached $ 592 billion in 1998, before more than doubling to $ 1.22 trillion in 1999--rivaling the US total. The global value of M&A activity in 1999 reached $ 3.4 trillion, an astounding 10% of world GDP. Next I will document that share issue privatizations have truly transformed share ownership patterns of investors in many different countries. B. Privatization's Impact on Stock and Bond Market Development We should be careful in inferring causation regarding privatization's impact on market growth, since a shift in ideology or some other exogenous political or economic change might have caused both the privatization and the overall boom. a. Total Proceeds Raised by Privatization Programs It is clear that national governments have been among the biggest winners from privatization programs, since these have dramatically increased government revenues, which is clearly one reason the policy has spread so rapidly. As mentioned above, Privatisation International [Gibbon (1998, 2000)] reports that the cumulative value of proceeds raised by privatizing governments exceeded $ 1 trillion sometime during the second half of 1999. As an added benefit, this revenue has come to governments without having to raise taxes or cut other public services. b. Privatization's Impact on International Investment Banking All international investment banks compete fiercely for share issue privatization mandates, for two principal reasons. First, because the offerings are so large and so visible--and are almost always designed to help promote the market's capacity to absorb subsequent stock offerings by private companies--these are very prestigious mandates. To date, the large US and British brokerage houses have had the most success in winning advisory and underwriting mandates, though all countries that launch large-scale SIP programs tend to favor local investment banks as "national champions" to handle the domestic share tranche. The second reason banks compete so fiercely for SIP mandates is because they can be extremely profitable. In spite of the fact--documented by Jones, et al (1999) and Ljungqvist, et al (2000)--that SIPs have significantly lower underwriting spreads than private sector offerings, their sheer size and lack of downside price risk make them very lucrative for underwriters. 2. Will this growth continue throughout the 2000s? As we indicated above, the global capital market has grown so rapidly in recent decades cause of the privatizations rise. Privatizations increased the market liquidity. Now we have already stepped into the 21st century. I believe that the growth will continue for the following reasons. First, most of the south-east Asia countries have recovered from the 1997 financial crisis. For these countries, they now have the capital to do businesses. And they get back on the fast growing track. Second, by the end of 2001, world's biggest developing country, China, has entered the WTO (World Trade Organization). This is real great news. As we all know, today's China takes a serious position in world's economy. Its innovation and opening policy make china keep achieving high GDP growth rate. This drives the global capital market keep growing. Summary and Conclusions This essay examines the impact of share issue privatizations (SIPs) on the growth of world capital markets (especially stock markets). I begin by documenting the increasing importance of capital markets, and the declining role of commercial banks, in corporate financial systems around the world. I then show that privatization programs-- particularly those involving public share offerings--have had a dramatic impact both on the development of non-US stock markets and on the participation of individual and institutional investors in those stock markets. This has told the reason of the fast growth of global capital market. And then I succinctly indicated the continuance of the rapid growth, the great future. The last but not the least is the recommendation. I can confidently assert that, if executed properly, a series of share issue privatizations can indeed promote the growth of global capital market, which will yield economic and political dividends for many years to come. That means there is a need to encourage the development of SIPs in order to gain growth of global capital market.
考点1:“market-based finance”推荐译为“市场主导式金融” 考点2:“The bits of banking” 推荐译为“银行业务” 考点3:“intermediary-based finance”应译为“以中介为基础的金融” 考点4:“syndicated credits”应译为“银团贷款” 考点5:“the weight of opinion” 应译为“主流观点” 考点6:“proceeds”应译为“单笔交易的收入所得” 考点7:“offerings”应译为“发行” 考点8:“prestigious mandates” 应译为“顶级授权” 考点9:“brokerage houses”应译为“证券经纪商” 考点10:“SIP”应译为“股票发行私有化”
08ce3
学术论文
社会科学
48
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 近年来,“替代两小时户外”的哺光仪、“模拟阳光”的大路灯等产品层出不穷。这些产品能否切实改善视力?背后是否暗藏风险?第30个全国“爱眼日”到来之际,听听权威专家的解读与建议。 三分钟哺光能否替代两小时户外? 哺光仪,也称重复低强度红光(RLRL),是一种以激光为光源照射眼睛,用于近视控制或弱视治疗的医疗设备产品。 记者在某电商平台上输入“哺光仪”搜索发现,相关产品品类众多,最高价格近4000元,有商家还提供800元/月的租赁使用服务。 2023年,因哺光仪不规范使用造成某12岁女童视网膜黄斑损伤,导致视力永久性受损。同年6月,国家药监局发布通知,将激光近视弱视治疗仪类产品划分为第三类医疗器械,并给予企业和市场一年过渡期。这意味着在2024年7月1日之后,企业生产、销售哺光仪,须具有第三类医疗器械注册证和生产许可证。 今年4月,北京大学人民医院、北京同仁医院有关专家在国际知名眼科期刊共同发表名为《近视儿童重复低强度红光治疗后视锥细胞密度的变化》的论文,指出以激光作为光源对儿童眼睛进行照射以防控近视,有引发视锥光感受器受损的风险。 中山大学孙逸仙纪念医院眼科副主任医师张一弛说,哺光仪目前临床研究观察最长时间为一年。部分孩子使用后眼轴增长确有所控制,但发生机制尚不明确,长期暴露情况下安全性、有效性也有待观察。 首都医科大学附属北京康复医院眼科主任刘莹说,部分家长因孩子近视进展或眼轴增长过快而选择使用哺光仪,也有一些家长认为孩子度数不严重,想用三分钟哺光替代两小时户外,这些想法并不可取。 这位专家表示,户外活动作为近视防控方案的循证医学证据在全球的观察时间更长、数据更多。建议家长们理性评估风险与收益,优先选择证据充分的防控手段。 大路灯能“模拟自然光”? “相比普通台灯,大路灯的室内照明光线分布相对均匀,能够减少阴影和暗区,照射范围也更广,在一定程度上有助于减轻视疲劳,但并不足以单独作为一种近视防控方法。”刘莹说,与按照三类医疗器械管理的哺光仪不同,大路灯本质上是一种灯具。 浙江大学眼科医院视光中心主任倪海龙表示,万物生长靠太阳,晴朗白天的太阳光照度可达10万勒克斯(lux),远超能提供1000lux左右光照强度的所谓大路灯。“同时,大路灯也无法替代户外光刺激视网膜分泌多巴胺的关键作用。”他说。 很多临床医生都在门诊中遇到家长请求推荐灯具产品。不少家长表示,大路灯“参数眼花缭乱、价格五花八门”,“缺乏行业标准、质量良莠不齐”,购买后发现部分低价产品夸大宣传。 《近视防治指南(2024年版)》明确,读写应在采光良好、照明充足的环境中进行,桌面的平均照度值不应低于300lux。国家标准《读写作业台灯性能要求》对灯具色温、显色指数、照度、视网膜蓝光危害和闪烁等多项指标提出要求。 专家强调,近视的发展受环境因素、遗传因素等共同影响,采光照明只是其中一个方面。选购灯具应优先参照国家出台的相关标准。 改善整体光环境和用眼习惯,避免依赖单一技术手段 倪海龙强调,近视防控的关键仍在于一增一减,即增加户外活动,减少近距离用眼负担,同时可辅以改善光环境及用眼习惯,要打组合拳,而非依赖单一技术手段。 专家建议要科学看待人工光源,辅助工具不可替代自然光照和基础防控。 “不论是儿童青少年还是成年人,自然光作为视觉的核心保护因素,其作用不可替代。”刘莹说,人工光源的应用需以“安全、循证”为原则,避免因商业营销或焦虑心理陷入误区。 近年来,党和国家高度重视青少年近视防控,部署三级干预措施。一级预防包括从婴幼儿期(2岁半起)就定期筛查视力,避免过早接触电子产品;二级预防包括通过户外活动、用眼习惯调整降低近视风险;三级预防包括采用离焦眼镜、角膜塑形镜及低浓度阿托品等医学手段延缓近视加深。 专家表示,近视防控是个系统工程,需要全社会行动起来,关注全生命周期的用眼健康。
近年来,“替代两小时户外”的哺光仪、“模拟阳光”的大路灯等产品层出不穷。这些产品能否切实改善视力?背后是否暗藏风险?第30个全国“爱眼日”到来之际,听听权威专家的解读与建议。 三分钟哺光能否替代两小时户外? 哺光仪,也称重复低强度红光(RLRL),是一种以激光为光源照射眼睛,用于近视控制或弱视治疗的医疗设备产品。 记者在某电商平台上输入“哺光仪”搜索发现,相关产品品类众多,最高价格近4000元,有商家还提供800元/月的租赁使用服务。 2023年,因哺光仪不规范使用造成某12岁女童视网膜黄斑损伤,导致视力永久性受损。同年6月,国家药监局发布通知,将激光近视弱视治疗仪类产品划分为第三类医疗器械,并给予企业和市场一年过渡期。这意味着在2024年7月1日之后,企业生产、销售哺光仪,须具有第三类医疗器械注册证和生产许可证。 今年4月,北京大学人民医院、北京同仁医院有关专家在国际知名眼科期刊共同发表名为《近视儿童重复低强度红光治疗后视锥细胞密度的变化》的论文,指出以激光作为光源对儿童眼睛进行照射以防控近视,有引发视锥光感受器受损的风险。 中山大学孙逸仙纪念医院眼科副主任医师张一弛说,哺光仪目前临床研究观察最长时间为一年。部分孩子使用后眼轴增长确有所控制,但发生机制尚不明确,长期暴露情况下安全性、有效性也有待观察。 首都医科大学附属北京康复医院眼科主任刘莹说,部分家长因孩子近视进展或眼轴增长过快而选择使用哺光仪,也有一些家长认为孩子度数不严重,想用三分钟哺光替代两小时户外,这些想法并不可取。 这位专家表示,户外活动作为近视防控方案的循证医学证据在全球的观察时间更长、数据更多。建议家长们理性评估风险与收益,优先选择证据充分的防控手段。 大路灯能“模拟自然光”? “相比普通台灯,大路灯的室内照明光线分布相对均匀,能够减少阴影和暗区,照射范围也更广,在一定程度上有助于减轻视疲劳,但并不足以单独作为一种近视防控方法。”刘莹说,与按照三类医疗器械管理的哺光仪不同,大路灯本质上是一种灯具。 浙江大学眼科医院视光中心主任倪海龙表示,万物生长靠太阳,晴朗白天的太阳光照度可达10万勒克斯(lux),远超能提供1000lux左右光照强度的所谓大路灯。“同时,大路灯也无法替代户外光刺激视网膜分泌多巴胺的关键作用。”他说。 很多临床医生都在门诊中遇到家长请求推荐灯具产品。不少家长表示,大路灯“参数眼花缭乱、价格五花八门”,“缺乏行业标准、质量良莠不齐”,购买后发现部分低价产品夸大宣传。 《近视防治指南(2024年版)》明确,读写应在采光良好、照明充足的环境中进行,桌面的平均照度值不应低于300lux。国家标准《读写作业台灯性能要求》对灯具色温、显色指数、照度、视网膜蓝光危害和闪烁等多项指标提出要求。 专家强调,近视的发展受环境因素、遗传因素等共同影响,采光照明只是其中一个方面。选购灯具应优先参照国家出台的相关标准。 改善整体光环境和用眼习惯,避免依赖单一技术手段 倪海龙强调,近视防控的关键仍在于一增一减,即增加户外活动,减少近距离用眼负担,同时可辅以改善光环境及用眼习惯,要打组合拳,而非依赖单一技术手段。 专家建议要科学看待人工光源,辅助工具不可替代自然光照和基础防控。 “不论是儿童青少年还是成年人,自然光作为视觉的核心保护因素,其作用不可替代。”刘莹说,人工光源的应用需以“安全、循证”为原则,避免因商业营销或焦虑心理陷入误区。 近年来,党和国家高度重视青少年近视防控,部署三级干预措施。一级预防包括从婴幼儿期(2岁半起)就定期筛查视力,避免过早接触电子产品;二级预防包括通过户外活动、用眼习惯调整降低近视风险;三级预防包括采用离焦眼镜、角膜塑形镜及低浓度阿托品等医学手段延缓近视加深。 专家表示,近视防控是个系统工程,需要全社会行动起来,关注全生命周期的用眼健康。
考点 1: "重复低强度红光(RLRL)" 应译为 "repeated low-level red light (RLRL)" 考点 2: "以激光为光源照射眼睛" 应译为 "laser-based ocular irradiation" 考点 3: "国家药监局" 应译为 "National Medical Products Administration (NMPA)" 考点 4: "注册证和生产许可证" 应译为 "registration certificate and production license" 考点 5: "视锥细胞密度" 应译为 "cone cell density" 考点 6: "视锥光感受器受损" 应译为 "cone photoreceptor damage" 考点 7: "质量良莠不齐" 应译为 "uneven product quality" 考点 8: "《近视防治指南(2024 年版)》" 应译为 "Guidelines for Myopia Prevention and Control (2024 edition)" 考点 9: "国家标准《读写作业台灯性能要求》" 应译为 "National Standard: Performance Requirements for Reading and Writing Desk Lamps" 考点 10: "一增一减" 应译为 "'one increase, one decrease"'strategy" 考点 11: "组合拳" 应译为 "a combination strategy" / "multi-pronged approach" 考点 12: "青少年近视防控" 应译为 "juvenile myopia prevention and control" 考点 13: "离焦眼镜" 应译为 "defocus spectacles" 考点14: “全国爱眼日”应译为“Sight Day”
0a0f2
新闻资讯
新闻报道
96
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 国家金融监督管理总局是中国负责金融行业统一监督管理的机构,主要职责包括强化机构监管、行为监管、功能监管等,以维护金融业的合法和稳健运行。该局还参与金融业改革开放和监管有效性相关问题的研究,并拟订相关法律法规草案。近期,该局还修订了《货币经纪公司管理办法》,并向社会公开征求意见,以加强对货币经纪公司的监管。此外,局长李云泽在新闻发布会上表示,将推出八项增量政策,以支持经济回升。 2025年6月,为加强商业银行的市场风险管理,国家金融监督管理总局根据《中华人民共和国银行业监督管理法》《中华人民共和国商业银行法》以及其他有关法律和行政法规,制定了《商业银行市场风险管理办法》,并印发给各金融监管局,各政策性银行、大型银行、股份制银行、外资银行、直销银行、金融资产管理公司、金融资产投资公司。 《商业银行市场风险管理办法》(下文简称《规则》)共五章四十三条,主要涵盖以下几个方面:首先,明确市场风险定义。《规则》明确了适用范围,明确排除了银行账簿中的利率风险,并加强了与《商业银行资本管理办法》和《银行业金融机构全面风险管理指引》的协调一致。其次,强调改进市场风险治理框架。规则阐明了董事会、监事会和高级管理层的责任,界定了三道防线具体范围和职责,并强调银行需要在集团层面加强市场风险管理。 第三,细化市场风险管理要求。规则要求银行进行端到端的市场风险管理,并详细规定了风险识别、计量、监测、控制和报告的总体要求。规则还改进了内部模型的定义,并提高了对模型治理和压力测试的要求,确保与当前市场风险计量框架和管理实践保持一致。未来,国家金融监督管理总局将加强监管和指导,确保规则的有效实施,并将指导银行提升其市场风险管理能力。 同时,国家金融监督管理总局还具有对于中国大陆境内金融机构评级的职责,例如,同样在2025年6月,为了积极参与国际保险集团的监管治理,维护全球金融市场的稳定,推动中国保险业高水平对外开放,国家金融监督管理总局根据国际保险监督协会发布的《国际保险集团监管共同框架》,对中国具有国际影响力的保险集团进行了评估和认定。经评估,中国再保险(集团)股份有限公司(简称“中国再保”)被认定为具有国际影响力的保险集团。今后,国家金融监督管理总局将指导中国再保按照《国际保险集团监管共同框架》进一步健全其风险管理体系,持续提升经营管理能力和国际竞争力。 提到再保险,可能很多人会觉得陌生。再保险也称分保,是保险人在原保险合同的基础上,通过签订分保合同,将其所承保的部分风险和责任向其他保险人进行保险的行为。再保险作为“保险的保险”,对于保障保险市场安全,为直接保险公司分散赔付风险、扩大承保能力和巨灾保障功能,并辅助保险市场调控以及强化行业风险管理发挥了重要的作用。以中国再保为例,中国再保险(集团)股份有限公司(简称“中国再保”)由中华人民共和国财政部和中央汇金投资有限责任公司发起设立,注册资本人民币42,479,808,085元,其中财政部持股11.45%,中央汇金投资有限责任公司持股71.56%。中国再保源于1949年10月成立的中国人民保险公司,2007年10月整体改制为股份有限公司。目前,中国再保主要控股7家境内子公司:中国财产再保险有限责任公司(简称“中再产险”)、中国人寿再保险有限责任公司(简称“中再寿险”)、中国大地财产保险股份有限公司等。 截止2024年11月,中国大陆境内15家再保险公司中有9家为外资再保险公司,分别是慕尼黑再保险公司北京分公司、德国通用再保险股份公司上海分公司、RGA美国再保险公司上海分公司、大韩再保险公司上海分公司、法国再保险公司北京分公司、汉诺威再保险股份公司上海分公司、瑞士再保险股份有限公司北京分公司、信利再保险(中国)有限公司、曼福再保险公司北京分公司;6家为内资机构分别是中国农业再保险股份有限公司、人保再保险股份有限公司、太平再保险(中国)有限公司、中国财产再保险有限责任公司、中国人寿再保险有限责任公司。 随着我国经济的快速增长和保险行业的不断发展,中国再保险市场专业主体逐渐丰富,保费规模和市场份额稳步扩大。但也要看到对标国际发达再保险市场水平,我国再保险市场还存在一定差距。业内人士建议,内资再保险公司可借助保险科技手段,加强自身的风险管理能力,更好地满足直保公司的服务需求。同时,加大资本投入,增强资本实力,提升自身的承保能力和抗风险能力。此外,加强国际合作与交流,学习国际先进的保险专业技术,利用国际市场进行风险分散,提升自己的承保能力与盈利能力。
国家金融监督管理总局是中国负责金融行业统一监督管理的机构,主要职责包括强化机构监管、行为监管、功能监管等,以维护金融业的合法和稳健运行。该局还参与金融业改革开放和监管有效性相关问题的研究,并拟订相关法律法规草案。近期,该局还修订了《货币经纪公司管理办法》,并向社会公开征求意见,以加强对货币经纪公司的监管。此外,局长李云泽在新闻发布会上表示,将推出八项增量政策,以支持经济回升。 2025年6月,为加强商业银行的市场风险管理,国家金融监督管理总局根据《中华人民共和国银行业监督管理法》《中华人民共和国商业银行法》以及其他有关法律和行政法规,制定了《商业银行市场风险管理办法》,并印发给各金融监管局,各政策性银行、大型银行、股份制银行、外资银行、直销银行、金融资产管理公司、金融资产投资公司。 《商业银行市场风险管理办法》(下文简称《规则》)共五章四十三条,主要涵盖以下几个方面:首先,明确市场风险定义。《规则》明确了适用范围,明确排除了银行账簿中的利率风险,并加强了与《商业银行资本管理办法》和《银行业金融机构全面风险管理指引》的协调一致。其次,强调改进市场风险治理框架。规则阐明了董事会、监事会和高级管理层的责任,界定了三道防线具体范围和职责,并强调银行需要在集团层面加强市场风险管理。 第三,细化市场风险管理要求。规则要求银行进行端到端的市场风险管理,并详细规定了风险识别、计量、监测、控制和报告的总体要求。规则还改进了内部模型的定义,并提高了对模型治理和压力测试的要求,确保与当前市场风险计量框架和管理实践保持一致。未来,国家金融监督管理总局将加强监管和指导,确保规则的有效实施,并将指导银行提升其市场风险管理能力。 同时,国家金融监督管理总局还具有对于中国大陆境内金融机构评级的职责,例如,同样在2025年6月,为了积极参与国际保险集团的监管治理,维护全球金融市场的稳定,推动中国保险业高水平对外开放,国家金融监督管理总局根据国际保险监督协会发布的《国际保险集团监管共同框架》,对中国具有国际影响力的保险集团进行了评估和认定。经评估,中国再保险(集团)股份有限公司(简称“中国再保”)被认定为具有国际影响力的保险集团。今后,国家金融监督管理总局将指导中国再保按照《国际保险集团监管共同框架》进一步健全其风险管理体系,持续提升经营管理能力和国际竞争力。 提到再保险,可能很多人会觉得陌生。再保险也称分保,是保险人在原保险合同的基础上,通过签订分保合同,将其所承保的部分风险和责任向其他保险人进行保险的行为。再保险作为“保险的保险”,对于保障保险市场安全,为直接保险公司分散赔付风险、扩大承保能力和巨灾保障功能,并辅助保险市场调控以及强化行业风险管理发挥了重要的作用。以中国再保为例,中国再保险(集团)股份有限公司(简称“中国再保”)由中华人民共和国财政部和中央汇金投资有限责任公司发起设立,注册资本人民币42,479,808,085元,其中财政部持股11.45%,中央汇金投资有限责任公司持股71.56%。中国再保源于1949年10月成立的中国人民保险公司,2007年10月整体改制为股份有限公司。目前,中国再保主要控股7家境内子公司:中国财产再保险有限责任公司(简称“中再产险”)、中国人寿再保险有限责任公司(简称“中再寿险”)、中国大地财产保险股份有限公司等。 截止2024年11月,中国大陆境内15家再保险公司中有9家为外资再保险公司,分别是慕尼黑再保险公司北京分公司、德国通用再保险股份公司上海分公司、RGA美国再保险公司上海分公司、大韩再保险公司上海分公司、法国再保险公司北京分公司、汉诺威再保险股份公司上海分公司、瑞士再保险股份有限公司北京分公司、信利再保险(中国)有限公司、曼福再保险公司北京分公司;6家为内资机构分别是中国农业再保险股份有限公司、人保再保险股份有限公司、太平再保险(中国)有限公司、中国财产再保险有限责任公司、中国人寿再保险有限责任公司。 随着我国经济的快速增长和保险行业的不断发展,中国再保险市场专业主体逐渐丰富,保费规模和市场份额稳步扩大。但也要看到对标国际发达再保险市场水平,我国再保险市场还存在一定差距。业内人士建议,内资再保险公司可借助保险科技手段,加强自身的风险管理能力,更好地满足直保公司的服务需求。同时,加大资本投入,增强资本实力,提升自身的承保能力和抗风险能力。此外,加强国际合作与交流,学习国际先进的保险专业技术,利用国际市场进行风险分散,提升自己的承保能力与盈利能力。
考点1:国家金融监督管理总局只能译为National Financial Regulatory Administration,因为这是固定机构名称 考点2:《中华人民共和国银行业监督管理法》只能为Banking Supervision Law of the People's Republic of China,因为这是法律文件名称 考点3:《中华人民共和国商业银行法》只能译为Law of the People's Republic of China on Commercial Banks,因为这是法律文件名称 考点4:《商业银行资本管理办法》只能译为Administrative Measures for the Capital of Commercial Banks,因为这是法律文件名称 考点5:《国际保险集团监管共同框架》只能译为Common Framework for the Supervision of Internationally Active Insurance Groups,因为这是法律文件名称 考点6:中央汇金投资有限责任公司只能译为Central Huijin Investment Ltd.,因为这是公司名称 考点7:中国人民保险公司只能译为The People's Insurance Company (Group) of China,因为这是公司名称 考点8:中国财产再保险有限责任公司(简称“中再产险”)只能译为China Property & Casualty Reinsurance Company Ltd.(CHINA RE P&C),因为这是公司名称 考点9:中国人寿再保险有限责任公司(简称“中再寿险”)只能译为China Life Reinsurance Company Ltd. (China Re Life),因为这是公司名称 考点10:中国大地财产保险股份有限公司只能译为China Continent Property & Casualty Insurance Co., Ltd (CCIC),因为这是公司名称 考点11:德国通用再保险股份公司上海分公司只能译为General Reinsurance AG Shanghai Branch,因为这是公司名称 考点12:RGA美国再保险公司上海分公司只能译为RGA Reinsurance Company Shanghai Branch,因为这是公司名称 考点13:大韩再保险公司上海分公司只能译为Korean Reinsurance Company, Shanghai Branch 或 Korean Reinsurance Company Shanghai Branch,因为这是公司名称 考点14:法国再保险公司北京分公司只能译为SCOR SE Beijing Branch,因为这是公司名称 考点15:汉诺威再保险股份公司上海分公司只能译为Hannover Rück SE Shanghai Branch,因为这是公司名称 考点16:瑞士再保险股份有限公司北京分公司只能译为Swiss Reinsurance Company Ltd. Beijing Branch,因为这是公司名称 考点17:曼福再保险公司北京分公司只能译为MAPFRE RE Beijing Branch,因为这是公司名称 考点18:中国农业再保险股份有限公司只能译为China Agricultural Reinsurance Co., Ltd.,因为这是公司名称 考点19:人保再保险股份有限公司只能译为PICC Reinsurance Company Limited,因为这是公司名称 考点20:太平再保险(中国)有限公司只能译为Taiping Reinsurance(China)Company Limited,因为这是公司名称
0c820
垂类场景
金融
198
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 与过去相比,如今我们与他人交往的方式真是空前多样。曾经,我们只能依靠面对面交谈,但在过去的几千年中,新的交往技术不断被创造出来。数字时代的独特之处,便在于使人与人相联系的技术中介经历了快速转型。在面对面交谈、固定电话、邮政信件等传统交往方式的基础上,我们拥有了电子邮件、移动电话、短信、即时通信、网聊、留言板、社交网络、照片分享、视频分享、多人在线游戏等诸多新型交往方式。不过,在面对新媒体时,人们也时常感到困惑。在这个创新与扩散日新月异的时代,我们自然会关心这些新型交往方式对人际关系的影响。 面对层出不穷的新媒体,我们往往有两种反应: 一些人对人际交往的浅薄化表示担忧,对于很多人而言,日益频繁的中介化互动似乎威胁到了人际关系的神圣性;另一些人则认为, 新媒体为我们创造了更多与他人建立联系的机会,从而形成了更强大、更多样化的关系链。这两种观点都有着深厚的文化历史背景,也都印证了同一种观念:数字媒体正在改变社会关系的本质。随着时间的推移,当我们逐渐习惯了新的传播媒体时,人们的态度开始发生微妙的改变。这些媒体的存在被视为理所当然,甚至可以忽略不计。所以,我们思考科技、探索交往,以及反思两者关联的最佳时刻,就是这些新媒体刚刚出现,有关它们的使用准则还未固定之时。 本书围绕数字媒体和数字设备在人际关系中扮演的角色展开,旨在为大家提供一种批判性思考的方式。比起目不暇接、发人深省的故事逸闻,本书更愿意提供一些理论和数据资料,帮助读者理解人际关系中发生的重要变化。我从1990年开始关注这个领域,1991年启动了我第一项有关网络人际关系的研究,1994年起在传播学院开设传播和新技术的相关课程。本书的素材取自我的研究项目、观察以及大量与此相关的其他研究文献,这些素材为评估和理解人际关系的变迁奠定了框架。 当我们试图理解数字媒体和它在我们生活中的位置,以及对我们的个性和人际关系的影响时,会发现各种各样值得思考的议题。在技术的最初发展阶段,它会影响我们如何看待世界、社区、关系和自我。这也会促使社会和文化的重构与反思。卡罗琳·马尔温的一项著名研究考察了19世纪大众科学杂志,她发现在人类历史中,电、电报、电话这些新技术的出现会将人们熟悉的事物陌生化,因此也更容易导致改变。这种改变又会造成人们的焦虑。在古代社会,人们曾为书写的出现担忧;在维多利亚时代,人们害怕电;如今,我们的“焦虑不仅针对电脑,还针对更广泛意义上的技术”。 从远古时代起,这些传播技术出现的根本目的就是能让人们在身体缺席时,仍旧能够传递信息。在19世纪电报发明之前,这种超越空间的能力不可避免地会伴随时间的延迟,信息传达给受众可能要用上几年的时间。随着电报的出现,人类在历史上首次实现了不受距离限制的实时通信。人们也许曾对写作和出版感到震惊,不过,与面临这种全新的、瓦解时空边界的力量时产生的震惊相比,前者只能算是小巫见大巫。毕竟数千年来,人类早已习惯面对面社交,这种以极高速度进行远距离交流的能力,打破了我们根植于集体意识深处的社会认知。数字媒体的出现则造成了更严重的困扰。它们向学者和普通人提出了许多重要的问题:怎样才能既在场又缺席?如果自我不再需要身体作为载体,它将是什么样子?我们为何会在拥有如此多控制权的同时,又丧失了如此多的自由?当个人交流通过大众媒体传播时,意味着什么?当大众传播被用于个人交流时,其到底该被如何定义?“私人”和“公共”如何区分?“真实”到底是什么意思? 肯尼思·格根认为我们正在与“缺席的在场所造成的挑战”做斗争。尽管在物理空间中,我们身边不乏有血有肉的人,但我们依然为自己在“漂浮世界”中与不在场的对象打交道感到忧心忡忡。雪莉·特克尔在《群体性孤独》一书中指出, 我们也许置身某地,但是心思和情感却在别处。例如,你的同伴虽然与你共进晚餐,但却一直低头用手机与别人聊天。那么他的身体既是在场的,但同时又是缺席的,于是,“自我”的本质就变得模糊起来。 “他”到底在哪里?哈拉维在《赛博格宣言》中宣告,人类和机器的界限已经瓦解。不仅如此,自我和身体的界限也陷入不确定性之中。很快,一些人便认为,他们的“真实自我”在网上能得到最佳呈现。异地恋的双方通过电子设备建立和维系关系,可穿戴设备被嵌入我们日常穿着之中,那么,我们又怎么确认真实的自我究竟在何处栖身呢?此外,如果数字媒体中的自我和面对面交流中的自我不一样,甚至相互矛盾时,我们又该怎么办呢?如果一个人在面对面交流中表现得很有教养,在一个网络论坛中咄咄逼人,在另一个论坛中则渴求关爱,那么,哪一个自我才是真实的呢?是否还存在真实的自我?它曾经存在过吗?
与过去相比,如今我们与他人交往的方式真是空前多样。曾经,我们只能依靠面对面交谈,但在过去的几千年中,新的交往技术不断被创造出来。数字时代的独特之处,便在于使人与人相联系的技术中介经历了快速转型。在面对面交谈、固定电话、邮政信件等传统交往方式的基础上,我们拥有了电子邮件、移动电话、短信、即时通信、网聊、留言板、社交网络、照片分享、视频分享、多人在线游戏等诸多新型交往方式。不过,在面对新媒体时,人们也时常感到困惑。在这个创新与扩散日新月异的时代,我们自然会关心这些新型交往方式对人际关系的影响。 面对层出不穷的新媒体,我们往往有两种反应: 一些人对人际交往的浅薄化表示担忧,对于很多人而言,日益频繁的中介化互动似乎威胁到了人际关系的神圣性;另一些人则认为, 新媒体为我们创造了更多与他人建立联系的机会,从而形成了更强大、更多样化的关系链。这两种观点都有着深厚的文化历史背景,也都印证了同一种观念:数字媒体正在改变社会关系的本质。随着时间的推移,当我们逐渐习惯了新的传播媒体时,人们的态度开始发生微妙的改变。这些媒体的存在被视为理所当然,甚至可以忽略不计。所以,我们思考科技、探索交往,以及反思两者关联的最佳时刻,就是这些新媒体刚刚出现,有关它们的使用准则还未固定之时。 本书围绕数字媒体和数字设备在人际关系中扮演的角色展开,旨在为大家提供一种批判性思考的方式。比起目不暇接、发人深省的故事逸闻,本书更愿意提供一些理论和数据资料,帮助读者理解人际关系中发生的重要变化。我从1990年开始关注这个领域,1991年启动了我第一项有关网络人际关系的研究,1994年起在传播学院开设传播和新技术的相关课程。本书的素材取自我的研究项目、观察以及大量与此相关的其他研究文献,这些素材为评估和理解人际关系的变迁奠定了框架。 当我们试图理解数字媒体和它在我们生活中的位置,以及对我们的个性和人际关系的影响时,会发现各种各样值得思考的议题。在技术的最初发展阶段,它会影响我们如何看待世界、社区、关系和自我。这也会促使社会和文化的重构与反思。卡罗琳·马尔温的一项著名研究考察了19世纪大众科学杂志,她发现在人类历史中,电、电报、电话这些新技术的出现会将人们熟悉的事物陌生化,因此也更容易导致改变。这种改变又会造成人们的焦虑。在古代社会,人们曾为书写的出现担忧;在维多利亚时代,人们害怕电;如今,我们的“焦虑不仅针对电脑,还针对更广泛意义上的技术”。 从远古时代起,这些传播技术出现的根本目的就是能让人们在身体缺席时,仍旧能够传递信息。在19世纪电报发明之前,这种超越空间的能力不可避免地会伴随时间的延迟,信息传达给受众可能要用上几年的时间。随着电报的出现,人类在历史上首次实现了不受距离限制的实时通信。人们也许曾对写作和出版感到震惊,不过,与面临这种全新的、瓦解时空边界的力量时产生的震惊相比,前者只能算是小巫见大巫。毕竟数千年来,人类早已习惯面对面社交,这种以极高速度进行远距离交流的能力,打破了我们根植于集体意识深处的社会认知。数字媒体的出现则造成了更严重的困扰。它们向学者和普通人提出了许多重要的问题:怎样才能既在场又缺席?如果自我不再需要身体作为载体,它将是什么样子?我们为何会在拥有如此多控制权的同时,又丧失了如此多的自由?当个人交流通过大众媒体传播时,意味着什么?当大众传播被用于个人交流时,其到底该被如何定义?“私人”和“公共”如何区分?“真实”到底是什么意思? 肯尼思·格根认为我们正在与“缺席的在场所造成的挑战”做斗争。尽管在物理空间中,我们身边不乏有血有肉的人,但我们依然为自己在“漂浮世界”中与不在场的对象打交道感到忧心忡忡。雪莉·特克尔在《群体性孤独》一书中指出, 我们也许置身某地,但是心思和情感却在别处。例如,你的同伴虽然与你共进晚餐,但却一直低头用手机与别人聊天。那么他的身体既是在场的,但同时又是缺席的,于是,“自我”的本质就变得模糊起来。 “他”到底在哪里?哈拉维在《赛博格宣言》中宣告,人类和机器的界限已经瓦解。不仅如此,自我和身体的界限也陷入不确定性之中。很快,一些人便认为,他们的“真实自我”在网上能得到最佳呈现。异地恋的双方通过电子设备建立和维系关系,可穿戴设备被嵌入我们日常穿着之中,那么,我们又怎么确认真实的自我究竟在何处栖身呢?此外,如果数字媒体中的自我和面对面交流中的自我不一样,甚至相互矛盾时,我们又该怎么办呢?如果一个人在面对面交流中表现得很有教养,在一个网络论坛中咄咄逼人,在另一个论坛中则渴求关爱,那么,哪一个自我才是真实的呢?是否还存在真实的自我?它曾经存在过吗?
考点1:“ 日新月异的时代”应该译为“ an era of rapid change”​,准确描述时代的特征 考点2:“何处栖身”推荐译为“where ···to dwell/reside”​
11af5
学术论文
社会科学
158
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务:  Abstract The landscape of mobile gaming has evolved significantly over the years, with profound changes in network reliability and traffic patterns. In the early 2010s, mobile games faced challenges due to unreliable networks and primarily featured asynchronous gameplay. However, in the current era, modern mobile games benefit from robust network connectivity, mirroring PC gaming experiences by relying on persistent connections to game servers. This shift prompted us to conduct an in-depth traffic analysis of two mobile games that represent opposite ends of the genre spectrum: a massively multiplayer game resembling PC MMORPGs with tightly synchronized gameplay, and a single-player puzzle game that incorporates asynchronous social interactions. Surprisingly, both games exhibited remarkably similar traffic footprints; small packets with short inter-packet arrival times, indicating their high expectations for network reliability. This suggests that game developers now prioritize network quality similarly to their PC gaming counterparts. Additionally, our analysis of packet lengths unveiled that recent mobile games predominantly employ short packets dominated by a few key packet types closely tied to player actions, which conforms to observations from PC online games. However, the self-similarity in traffic patterns, a notable feature in PC online games, only partially explains the traffic in mobile games, varying across genres. These findings shed light on the evolving traffic patterns in mobile games and emphasize the need for further research in this dynamic domain. Keywords: mobile games; traffic analysis; Internet measurement 1. Introduction Network traffic analysis reports have illustrated that gaming is one of the most popular Internet applications, accounting for an estimated 8–10% of total Internet traffic and ranking as the third biggest traffic source after video streaming and web applications, including social media [1,2,3]. Another report highlights that 92.3% of Internet users access the Internet using mobile phones, with gaming being the most common use for these devices [4]. Given that game servers are typically hosted in cloud data centers and that game service providers incur significant costs for network traffic, understanding mobile game traffic patterns is crucial, not only from an engineering viewpoint but also from a business perspective. As mobile devices, including smartphones and tablets, have become the prevailing service environment for the gaming industry, game genres that run on mobile devices have also become diverse and complicated. In the early era of networked mobile gaming in the 2010s, most mobile games were asynchronous due to the poor robustness and high cost of mobile networks. In these games, players could interact with other players, such as family members and friends, but they generally played independently rather than tightly together. On the other hand, today’s mobile network technology, especially in metropolitan areas, has significantly improved in connection stability. A recent survey showed that from October 2022 through March 2023, 5G and 4G networks in the UK exhibited 98.4% and 97.8% connection success rates on average, respectively, when a mobile device becomes active [5]. This suggests that mobile games enjoy much more robust network connectivity, even in the mobile network, and thus such games now rely more on a persistent network connection to the game server. This phenomenal change suggests that the current underlying network traffic patterns might differ from those of the 2010s and from those in the PC gaming environment. This paper illustrates traffic analysis from two mobile games with global service. The analyzed games represent opposite extremes in terms of their genre. One is strongly synchronized and a massive multiplayer game similar to that of a PC environment, but with an “auto-hunt” feature to allow gameplay without human engagement. The other is less synchronized and primarily played by a single player, although it also features social interactions among players. The primary contribution of our work is demonstrating that recent mobile games exhibit traffic patterns akin to those in PC games, particularly regarding packet length and inter-packet arrival times, but have differences regarding traffic self-similarity. Our analysis of the games’ traffic traces reveals that current mobile games typically feature notably short inter-packet arrival times, similar to traditional multiplayer games in the PC environment, accompanied by a long-tail distribution due to the intermittent sleep and resume nature of mobile applications. Concerning packet lengths, recent mobile games predominantly use shorter packets, favoring them over aggregated larger packets to conserve network bandwidth. These patterns, observed consistently across different game genres, suggest that modern mobile games are developed with expectations of reliable network connectivity, a notable shift from earlier in the 2010s. This consistency in traffic characteristics, irrespective of game genre, underscores a fundamental similarity in network usage between contemporary mobile and PC gaming platforms. Our analysis, however, discovered that the presence of self-similarity in traffic patterns, identified by previous research on PC online games’ traffic patterns, varies across game genres. This finding encourages further research to model mobile game traffic. This paper is organized as follows: Section 2 reviews research related to the mobile gaming industry. Section 3 briefly introduces the mobile games analyzed and explains the methods for collecting and anonymizing traffic data. In Section 4, we describe our analysis methodology and present key findings. Finally, Section 5 summarizes these findings and proposes directions for future research. 2. Related Work Reports have shown that gaming is the most popular activity among mobile device users. Furthermore, gaming activities, which include those on PCs, mobile devices, and consoles, collectively constitute the third largest source of total Internet traffic [1,2,3,4]. Despite its significance in the context of today’s Internet traffic, there is a notable dearth of comprehensive research on mobile game traffic. In contrast, considerable research efforts have been directed towards analyzing PC online game traces, modeling traffic patterns, and optimizing traffic [6,7,8,9,10,11,12,13]. Previous research agrees that PC online games generate highly periodic bursty short packets. In particular, Chen et al. discovered traffic exhibits pronounced periodicity and temporal locality in inter-packet arrival times, attributable to player action patterns by a comprehensive traffic analysis using substantial packet traces from a PC MMORPG game [11]. Feng et al. also observed that the distribution of game session time in PC games is not heavy-tailed, a characteristic stemming from the synchronized nature of gameplay [12]. Henderson et al. explored the network quality of service (QoS) tolerance of game players [14]. Apart from previous research mainly focusing on PC online games, our work fills the gap for mobile games. We confirmed that mobile games share similarities with PC online games regarding traffic patterns while having unique differences at the same time. Chen et al. conducted a preliminary traffic analysis using Pokémon Go, which is a mobile AR game [15], but their preliminary results did not uncover relationships with PC online games. There is considerable research measuring the real-life performance of various mobile network technologies (i.e., the 3G, 4G, and 5G) [5,16,17,18,19,20,21,22,23]. Based on these research findings, mobile application developers can reasonably assume that even 4G LTE, currently the most prevalent mobile network worldwide, can provide an average round-trip time (RTT) of about 50 ms between a carrier network and a player’s device in metro areas. In particular, Ref. [22] reports that a study conducted in London showed a 5-millisecond latency could be achieved with 99.999% reliability over a 5G network. This research indicates that applications, including games, running on mobile networks can reliably expect lower latency as mobile infrastructure evolves. The finding in mobile network robustness is the main motivation for our research on mobile game traffic patterns. Kämäräinen et al. investigated factors contributing to end-to-end latency in cloud gaming, where game scenes are rendered at cloud servers. Their research highlights emerging technologies that could enhance gaming experiences demanding high network bandwidth and strict latency constraints [24]. Similarly, Braud et al. conducted research on mobile AR applications, which require computational offloading to cloud servers and are, therefore, also subject to network bandwidth and latency limitations [25]. Despite the disruptive approach of cloud gaming, it was not widely deployed in the mobile gaming industry because of high costs and the limited robustness of mobile networks. This approach is, however, being revitalized by the rapid emergence of virtual reality technologies and technological improvements in mobile networks. Therefore, understanding mobile traffic characteristics through our research can also contribute to cloud gaming. 3. Materials and Methods Due to the exclusivity or closed nature of the gaming industry, game traffic traces are not readily accessible, even in anonymized form. Game companies hesitate to release traffic traces not only because of concerns over players’ privacy information but also due to fears that the data might be used to compromise the security of their services. This partially explains why there has been limited research conducted on the network traffic characteristics of online/mobile games, despite their significance in terms of traffic usage. For this limitation, we selected two representative game genres at opposite ends of the spectrum and collected network traffic traces from a mobile game for each genre. The games are globally serviced and independently developed and operated by different game companies. Therefore, the games do not share any specific development methodology or assumptions about underlying network behavior.
 Abstract The landscape of mobile gaming has evolved significantly over the years, with profound changes in network reliability and traffic patterns. In the early 2010s, mobile games faced challenges due to unreliable networks and primarily featured asynchronous gameplay. However, in the current era, modern mobile games benefit from robust network connectivity, mirroring PC gaming experiences by relying on persistent connections to game servers. This shift prompted us to conduct an in-depth traffic analysis of two mobile games that represent opposite ends of the genre spectrum: a massively multiplayer game resembling PC MMORPGs with tightly synchronized gameplay, and a single-player puzzle game that incorporates asynchronous social interactions. Surprisingly, both games exhibited remarkably similar traffic footprints; small packets with short inter-packet arrival times, indicating their high expectations for network reliability. This suggests that game developers now prioritize network quality similarly to their PC gaming counterparts. Additionally, our analysis of packet lengths unveiled that recent mobile games predominantly employ short packets dominated by a few key packet types closely tied to player actions, which conforms to observations from PC online games. However, the self-similarity in traffic patterns, a notable feature in PC online games, only partially explains the traffic in mobile games, varying across genres. These findings shed light on the evolving traffic patterns in mobile games and emphasize the need for further research in this dynamic domain. Keywords: mobile games; traffic analysis; Internet measurement 1. Introduction Network traffic analysis reports have illustrated that gaming is one of the most popular Internet applications, accounting for an estimated 8–10% of total Internet traffic and ranking as the third biggest traffic source after video streaming and web applications, including social media [1,2,3]. Another report highlights that 92.3% of Internet users access the Internet using mobile phones, with gaming being the most common use for these devices [4]. Given that game servers are typically hosted in cloud data centers and that game service providers incur significant costs for network traffic, understanding mobile game traffic patterns is crucial, not only from an engineering viewpoint but also from a business perspective. As mobile devices, including smartphones and tablets, have become the prevailing service environment for the gaming industry, game genres that run on mobile devices have also become diverse and complicated. In the early era of networked mobile gaming in the 2010s, most mobile games were asynchronous due to the poor robustness and high cost of mobile networks. In these games, players could interact with other players, such as family members and friends, but they generally played independently rather than tightly together. On the other hand, today’s mobile network technology, especially in metropolitan areas, has significantly improved in connection stability. A recent survey showed that from October 2022 through March 2023, 5G and 4G networks in the UK exhibited 98.4% and 97.8% connection success rates on average, respectively, when a mobile device becomes active [5]. This suggests that mobile games enjoy much more robust network connectivity, even in the mobile network, and thus such games now rely more on a persistent network connection to the game server. This phenomenal change suggests that the current underlying network traffic patterns might differ from those of the 2010s and from those in the PC gaming environment. This paper illustrates traffic analysis from two mobile games with global service. The analyzed games represent opposite extremes in terms of their genre. One is strongly synchronized and a massive multiplayer game similar to that of a PC environment, but with an “auto-hunt” feature to allow gameplay without human engagement. The other is less synchronized and primarily played by a single player, although it also features social interactions among players. The primary contribution of our work is demonstrating that recent mobile games exhibit traffic patterns akin to those in PC games, particularly regarding packet length and inter-packet arrival times, but have differences regarding traffic self-similarity. Our analysis of the games’ traffic traces reveals that current mobile games typically feature notably short inter-packet arrival times, similar to traditional multiplayer games in the PC environment, accompanied by a long-tail distribution due to the intermittent sleep and resume nature of mobile applications. Concerning packet lengths, recent mobile games predominantly use shorter packets, favoring them over aggregated larger packets to conserve network bandwidth. These patterns, observed consistently across different game genres, suggest that modern mobile games are developed with expectations of reliable network connectivity, a notable shift from earlier in the 2010s. This consistency in traffic characteristics, irrespective of game genre, underscores a fundamental similarity in network usage between contemporary mobile and PC gaming platforms. Our analysis, however, discovered that the presence of self-similarity in traffic patterns, identified by previous research on PC online games’ traffic patterns, varies across game genres. This finding encourages further research to model mobile game traffic. This paper is organized as follows: Section 2 reviews research related to the mobile gaming industry. Section 3 briefly introduces the mobile games analyzed and explains the methods for collecting and anonymizing traffic data. In Section 4, we describe our analysis methodology and present key findings. Finally, Section 5 summarizes these findings and proposes directions for future research. 2. Related Work Reports have shown that gaming is the most popular activity among mobile device users. Furthermore, gaming activities, which include those on PCs, mobile devices, and consoles, collectively constitute the third largest source of total Internet traffic [1,2,3,4]. Despite its significance in the context of today’s Internet traffic, there is a notable dearth of comprehensive research on mobile game traffic. In contrast, considerable research efforts have been directed towards analyzing PC online game traces, modeling traffic patterns, and optimizing traffic [6,7,8,9,10,11,12,13]. Previous research agrees that PC online games generate highly periodic bursty short packets. In particular, Chen et al. discovered traffic exhibits pronounced periodicity and temporal locality in inter-packet arrival times, attributable to player action patterns by a comprehensive traffic analysis using substantial packet traces from a PC MMORPG game [11]. Feng et al. also observed that the distribution of game session time in PC games is not heavy-tailed, a characteristic stemming from the synchronized nature of gameplay [12]. Henderson et al. explored the network quality of service (QoS) tolerance of game players [14]. Apart from previous research mainly focusing on PC online games, our work fills the gap for mobile games. We confirmed that mobile games share similarities with PC online games regarding traffic patterns while having unique differences at the same time. Chen et al. conducted a preliminary traffic analysis using Pokémon Go, which is a mobile AR game [15], but their preliminary results did not uncover relationships with PC online games. There is considerable research measuring the real-life performance of various mobile network technologies (i.e., the 3G, 4G, and 5G) [5,16,17,18,19,20,21,22,23]. Based on these research findings, mobile application developers can reasonably assume that even 4G LTE, currently the most prevalent mobile network worldwide, can provide an average round-trip time (RTT) of about 50 ms between a carrier network and a player’s device in metro areas. In particular, Ref. [22] reports that a study conducted in London showed a 5-millisecond latency could be achieved with 99.999% reliability over a 5G network. This research indicates that applications, including games, running on mobile networks can reliably expect lower latency as mobile infrastructure evolves. The finding in mobile network robustness is the main motivation for our research on mobile game traffic patterns. Kämäräinen et al. investigated factors contributing to end-to-end latency in cloud gaming, where game scenes are rendered at cloud servers. Their research highlights emerging technologies that could enhance gaming experiences demanding high network bandwidth and strict latency constraints [24]. Similarly, Braud et al. conducted research on mobile AR applications, which require computational offloading to cloud servers and are, therefore, also subject to network bandwidth and latency limitations [25]. Despite the disruptive approach of cloud gaming, it was not widely deployed in the mobile gaming industry because of high costs and the limited robustness of mobile networks. This approach is, however, being revitalized by the rapid emergence of virtual reality technologies and technological improvements in mobile networks. Therefore, understanding mobile traffic characteristics through our research can also contribute to cloud gaming. 3. Materials and Methods Due to the exclusivity or closed nature of the gaming industry, game traffic traces are not readily accessible, even in anonymized form. Game companies hesitate to release traffic traces not only because of concerns over players’ privacy information but also due to fears that the data might be used to compromise the security of their services. This partially explains why there has been limited research conducted on the network traffic characteristics of online/mobile games, despite their significance in terms of traffic usage. For this limitation, we selected two representative game genres at opposite ends of the spectrum and collected network traffic traces from a mobile game for each genre. The games are globally serviced and independently developed and operated by different game companies. Therefore, the games do not share any specific development methodology or assumptions about underlying network behavior.
考点 1:“traffic patterns” 应译为 “流量模式 / 流量特征” 考点 2:“asynchronous gameplay” 应译为 ”异步游戏玩法” 考点 3:“persistent connections” 应译为 “持久连接” 考点 4:“asynchronous social interactions 应译为 “异步社交互动” 考点 6:“traffic footprints” 应译为 “ 流量足迹” 考点 7:“inter-packet arrival times” 应译为 “分组间到达时间“ 考点 8:“packet lengths” 应译为 ”数据包长度“ 考点 9:“mobile devices” 应译为 “移动设备” 考点 10:“persistent network connection” 应译为 “持续网络连接“ 考点 11:”traffic traces“ 应译为 ”流量轨迹 / 流量跟踪数据“ 考点 12:“long-tail distribution” 应译为 “长尾分布” 考点 13:“game session time“ 应译为 “游戏会话时长“ 考点 14:”Quality of Service (QoS)“ 应译为 ”服务质量“ 考点 15:“packet traces” 应译为 “数据包跟踪数据” 考点 16:“Pokémon Go“ 应译为 “精灵宝可梦 Go“ 考点 17:”round-trip time (RTT)“ 应译为 ”往返时延“ 考点 18:"latency" 应译为 “时延 / 延迟” 考点 19:"cloud gaming" 应译为 “云游戏“ 考点 20:"end-to-end latency" 应译为 "端到端时延" 考点 21:"computational offloading" 应译为 "计算卸载" 考点 22:"network traffic characteristics" 应译为 "网络流量特征"
11dca
学术论文
应用学科
81
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 药品专利链接制度中的利益平衡与纠纷解决机制 一、专利链接的核心机制与运行冲突 药品专利链接制度(即 “药品上市审批与专利纠纷早期解决挂钩”)是平衡创新药企业与仿制药企业利益的关键制度。根据《药品专利纠纷早期解决机制实施办法》,仿制药申请人在提交上市申请时,需对参比制剂(即原研药)的专利状态作出声明,分为四类: 1. 专利无效声明(声明原研药专利全部无效) 2. 不侵权声明(声明仿制药技术方案未落入原研药专利保护范围) 3. 等待声明(声明在原研药专利到期后再上市) 4. 专利挑战声明(声明原研药专利应当被宣告无效或仿制药不侵权,并启动纠纷解决程序) 实践中,该机制的运行冲突集中在三方面:一是 “专利常青” 问题,原研药企业通过 “晶型专利 + 适应症专利” 的组合延长保护期(如某抗癌药核心专利到期后,企业通过新晶型专利再获 8 年保护),仿制药企业认为此举超出合理保护范围;二是 “首仿药市场独占期” 争夺,根据规定,首个成功挑战专利并获批的仿制药可获 12 个月市场独占期,但 2023 年某抗生素仿制药案中,两家企业同时挑战成功,引发 “谁应享有独占期” 的争议;三是 “审批周期与诉讼时效的衔接”,仿制药上市审批平均需 10 个月,而专利诉讼一审周期约 6 个月,可能出现 “药已上市但专利纠纷未决” 的情况,导致侵权风险。 二、利益平衡的实践困境 制度设计的初衷是 “激励创新与促进可及性”,但实践中面临三重矛盾: 1. 专利保护强度与药品可及性的冲突 创新药企业主张 “严格专利保护是研发动力”(某单抗药物研发成本超 10 亿美元,专利期内需收回成本),而患者组织则认为 “过长保护期推高药价”(某乙肝新药年治疗费 2.8 万元,专利到期后仿制药价格降至 3000 元)。2024 年某罕见病药案中,原研药专利保护期还有 5 年,但国内患者年死亡率达 30%,仿制药企业申请 “专利强制许可” 被驳回,引发 “生命权与专利权孰先” 的讨论。 2. 数据独占与试验数据依赖的矛盾 原研药企业享有 6 年临床试验数据独占期,仿制药企业需自行开展试验或寻求授权,但复杂制剂(如缓释微球)的试验数据难以重复,导致仿制药研发成本增加。某降糖药仿制药企业因无法获取原研药的药代动力学数据,试验周期延长 2 年,上市时间滞后于国际市场,被质疑 “数据独占变相延长专利保护”。 3. 跨境药品贸易中的平行进口争议 专利链接制度仅适用于境内上市药品,境外已上市但未在我国注册的原研药通过平行进口进入国内时,其专利状态不受该机制约束。2023 年某抗艾滋病药平行进口案中,境外低价药因未在我国登记专利,仿制药企业无法发起专利挑战,导致 “同药不同价”(进口价为国内仿制药的 1/3),冲击国内市场秩序。 三、纠纷解决路径的选择与局限 现行机制提供三类纠纷解决途径,但各有不足: 1. 行政裁决的效率优势与效力局限 国家药监局下设的专利纠纷早期解决机制办公室可在 90 日内作出行政裁决,效率高于诉讼,但裁决仅对当事人具有约束力,且不能直接宣告专利无效(需另行向专利局提出无效宣告请求)。2023 年某降压药案中,行政裁决认定 “仿制药不侵权”,但原研药企业不服并提起民事诉讼,导致同一纠纷重复处理。 2. 民事诉讼的终局性与周期问题 法院审理可直接对专利有效性作出认定(根据《专利法》第 76 条),但一审平均周期 8 个月,远超仿制药审批周期。某抗病毒药仿制药在诉讼期间获批上市,后法院认定侵权,企业被迫召回已销售药品,损失超 5000 万元。 3. 调解机制的自愿性与执行短板 行业协会组织的调解可快速达成和解(如 2024 年某抗生素案调解耗时 45 天),但和解协议需依赖双方自觉履行,缺乏强制执行力。某仿制药企业与原研药企业达成 “专利许可费 3%” 的调解协议后,仿制药上市后拒绝支付,原研药企业仍需通过诉讼维权。 四、制度完善的实践探索 针对上述问题,实践中已出现三类改进方向: 1. 专利信息登记的精细化 药监局建立 “药品专利信息登记平台”,要求原研药企业按 “活性成分、晶型、制剂、适应症” 分类登记专利,避免 “模糊登记”。某生物制剂企业因将 “通用技术特征” 登记为专利,被驳回登记请求,减少了 “专利流氓” 的投机空间。 2. 首仿药独占期的细化规则 明确 “同时挑战成功时,按首仿药申请日期排序”,2024 年某抗肿瘤仿制药案中,两家企业同日挑战成功,按 “技术先进性”(如生物利用度更高)判定其中一家享有独占期,减少了争议。 3. 跨境协作机制的试点 与东南亚国家建立 “药品专利数据互认”,某仿制药企业在泰国获批的生物等效性数据,在国内申报时被认可,缩短研发周期 6 个月,同时通过 “专利审查高速路”(PPH)加快跨境专利纠纷解决,2024 年某抗生素案通过 PPH 机制使中美专利审查周期同步,避免了重复诉讼。
药品专利链接制度中的利益平衡与纠纷解决机制 一、专利链接的核心机制与运行冲突 药品专利链接制度(即 “药品上市审批与专利纠纷早期解决挂钩”)是平衡创新药企业与仿制药企业利益的关键制度。根据《药品专利纠纷早期解决机制实施办法》,仿制药申请人在提交上市申请时,需对参比制剂(即原研药)的专利状态作出声明,分为四类: 1. 专利无效声明(声明原研药专利全部无效) 2. 不侵权声明(声明仿制药技术方案未落入原研药专利保护范围) 3. 等待声明(声明在原研药专利到期后再上市) 4. 专利挑战声明(声明原研药专利应当被宣告无效或仿制药不侵权,并启动纠纷解决程序) 实践中,该机制的运行冲突集中在三方面:一是 “专利常青” 问题,原研药企业通过 “晶型专利 + 适应症专利” 的组合延长保护期(如某抗癌药核心专利到期后,企业通过新晶型专利再获 8 年保护),仿制药企业认为此举超出合理保护范围;二是 “首仿药市场独占期” 争夺,根据规定,首个成功挑战专利并获批的仿制药可获 12 个月市场独占期,但 2023 年某抗生素仿制药案中,两家企业同时挑战成功,引发 “谁应享有独占期” 的争议;三是 “审批周期与诉讼时效的衔接”,仿制药上市审批平均需 10 个月,而专利诉讼一审周期约 6 个月,可能出现 “药已上市但专利纠纷未决” 的情况,导致侵权风险。 二、利益平衡的实践困境 制度设计的初衷是 “激励创新与促进可及性”,但实践中面临三重矛盾: 1. 专利保护强度与药品可及性的冲突 创新药企业主张 “严格专利保护是研发动力”(某单抗药物研发成本超 10 亿美元,专利期内需收回成本),而患者组织则认为 “过长保护期推高药价”(某乙肝新药年治疗费 2.8 万元,专利到期后仿制药价格降至 3000 元)。2024 年某罕见病药案中,原研药专利保护期还有 5 年,但国内患者年死亡率达 30%,仿制药企业申请 “专利强制许可” 被驳回,引发 “生命权与专利权孰先” 的讨论。 2. 数据独占与试验数据依赖的矛盾 原研药企业享有 6 年临床试验数据独占期,仿制药企业需自行开展试验或寻求授权,但复杂制剂(如缓释微球)的试验数据难以重复,导致仿制药研发成本增加。某降糖药仿制药企业因无法获取原研药的药代动力学数据,试验周期延长 2 年,上市时间滞后于国际市场,被质疑 “数据独占变相延长专利保护”。 3. 跨境药品贸易中的平行进口争议 专利链接制度仅适用于境内上市药品,境外已上市但未在我国注册的原研药通过平行进口进入国内时,其专利状态不受该机制约束。2023 年某抗艾滋病药平行进口案中,境外低价药因未在我国登记专利,仿制药企业无法发起专利挑战,导致 “同药不同价”(进口价为国内仿制药的 1/3),冲击国内市场秩序。 三、纠纷解决路径的选择与局限 现行机制提供三类纠纷解决途径,但各有不足: 1. 行政裁决的效率优势与效力局限 国家药监局下设的专利纠纷早期解决机制办公室可在 90 日内作出行政裁决,效率高于诉讼,但裁决仅对当事人具有约束力,且不能直接宣告专利无效(需另行向专利局提出无效宣告请求)。2023 年某降压药案中,行政裁决认定 “仿制药不侵权”,但原研药企业不服并提起民事诉讼,导致同一纠纷重复处理。 2. 民事诉讼的终局性与周期问题 法院审理可直接对专利有效性作出认定(根据《专利法》第 76 条),但一审平均周期 8 个月,远超仿制药审批周期。某抗病毒药仿制药在诉讼期间获批上市,后法院认定侵权,企业被迫召回已销售药品,损失超 5000 万元。 3. 调解机制的自愿性与执行短板 行业协会组织的调解可快速达成和解(如 2024 年某抗生素案调解耗时 45 天),但和解协议需依赖双方自觉履行,缺乏强制执行力。某仿制药企业与原研药企业达成 “专利许可费 3%” 的调解协议后,仿制药上市后拒绝支付,原研药企业仍需通过诉讼维权。 四、制度完善的实践探索 针对上述问题,实践中已出现三类改进方向: 1. 专利信息登记的精细化 药监局建立 “药品专利信息登记平台”,要求原研药企业按 “活性成分、晶型、制剂、适应症” 分类登记专利,避免 “模糊登记”。某生物制剂企业因将 “通用技术特征” 登记为专利,被驳回登记请求,减少了 “专利流氓” 的投机空间。 2. 首仿药独占期的细化规则 明确 “同时挑战成功时,按首仿药申请日期排序”,2024 年某抗肿瘤仿制药案中,两家企业同日挑战成功,按 “技术先进性”(如生物利用度更高)判定其中一家享有独占期,减少了争议。 3. 跨境协作机制的试点 与东南亚国家建立 “药品专利数据互认”,某仿制药企业在泰国获批的生物等效性数据,在国内申报时被认可,缩短研发周期 6 个月,同时通过 “专利审查高速路”(PPH)加快跨境专利纠纷解决,2024 年某抗生素案通过 PPH 机制使中美专利审查周期同步,避免了重复诉讼。
考点1:“参比制剂” 推荐译为 “reference listed drug (RLD)” 考点2:“首仿药市场独占期” 推荐译为 “first generic drug market exclusivity period” 考点3:“平行进口” 推荐译为 “parallel importation” 考点4:“晶型专利” 推荐译为 “crystal form patent” 考点5:“参比制剂” 必须译为 “reference listed drug (RLD)”,不能译为 “reference drug”
1271a
学术论文
社会科学
148
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: For every spike in voltage there was a small but predictable increase in pleasure With so much variety, it is telling when something remains constant. Try an experiment: lick your fingers as though you were about to turn a page. Instinctively, you’ve licked the spot where fingers grip light objects, and at its centre are the concentric ridges and grooves that define your fingerprint. If you move your finger over an object in most directions, the object will run roughly perpendicular to these ridges, allowing friction to tug on each ridge as though toppling a wall. This central, bulbous part of your fingertip also contains the finest, densest set of ridges. You can see this if you follow your finger a short distance toward your palm, where the ridges become progressively wider. It is no coincidence that the ridges are finest, most centred on the part of your finger that first makes contact with an object. It is also where the nerve endings that sense touch are most dense. If you’re the caressing sort, recall how you have touched a lover, your fingertips scanning as they glide slowly over skin. Perhaps your palm lay flat, presenting the largest possible surface for contact. The ridges of our fingers and hands are densely innervated by sensory neurons, nerve cells that translate pressure into changes in voltage. These sensory neurons come in a variety of forms suited for their tasks, named after neuroscientists like Merkel, Ruffini, Meissner and Pacini. Nerve endings can be capped with structures called disks, capsules or corpuscles – each defined by a distinctive weight or stiffness. These tips make the neurons more or less sensitive to pressure. The nerve endings that sense touch can be buried deep in the skin or can be so near the surface you could find them within the ridge of a fingerprint. When the pressure and depth of touch are just right, the surface of the sensing neuron is deformed, stretched until the tension opens channels that let electrically charged salt ions flow in and out of the cell. The voltage change caused by the flow of ions zips along a cable-like projection to the spinal cord, where it gets passed on to other nerve cells and eventually to the brain. We can judge how smooth or pliant something is because voltages conveying the complex patterns of pressure arrive quickly enough for our brains to perceive subtle variation in timing. Without this ability, touch would feel like a surveillance tape played at half-speed: blurred and coarse. Like other species, we gain this speed by insulating our cables. Nerve cells are highly specialised, and require companion cells to help them with the daily details of cellular living. Some of these companions have developed means of enveloping the cable-like projections of neurons, becoming flat and wrapping themselves around the exterior of the cable again and again, like a king-sized sheet swaddling an infant. Or like rubber coating wire. Insulated neurons are responsible for fine touch, but there is a second class of receptors that remain bare. These bare nerve endings are slower, and respond to coarser kinds of stimuli. Science has long known that these unmyelinated neurons respond to temperature, pain, tickle and itch. But we have only recently learned that they also respond to the pleasurable sensation of caress. Researchers in Sweden recorded data from neurons in the skin of human subjects as they exposed them to soft slow touch. For every spike in voltage there was a small but predictable increase in pleasure. While these naked neurons are missing in the hairless skin of our fingers and palms, they are found on the rest of the body, on the places you might touch with affection or consolation. And naked fibres are particularly abundant in the places we like to juxtapose – our lips, nipples, genitals and anus. The clitoris and the glans are enmeshed in the unmyelinated ends of sensory neurons. Inexplicably, we have often assumed these naked fibres were there for the sensation of pain, as though we had never known the joy of sexual touch. n the naked stream, touch can be warm or rapturous or full of hurt Each touch receptor propagates voltages upward toward the spinal cord and brain, voltages that float like bottles bearing notes along a waterway defined by the spindly extensions of sensory neurons. Each current conveys its own kind of message, and the myriad currents coalesce into two north-bound streams. Of these streams, the routes of discriminative touch are particularly well mapped. In the 1930s, Canadian neurosurgeon Wilder Penfield electrically stimulated the brains of epileptics, probing the cortex for the origin of seizures. Patients had to be awake for this procedure so that he could ask them what experiences were evoked by the faint electrical current. Electricity alone was enough to elicit the feeling of being touched on an arm, or, when delivered to a nearby region of cortex, the shoulder. Penfield found that the brain contained precise maps of the body; he charted duplicate maps of both touch and movement, side by side, along adjacent folds of the cortex. The resulting ‘homunculus’ is an iconic image in neuroscience – a strange representation of the body whose distortions, like early maps of the world, reflect how we value the body’s surface. Those areas where touch is most sensitive are inflated. And three-dimensional reconstructions of these maps reveal a grotesque caricature of our evolutionary past. Our fingers, faces, palms, lips, tongues and genitals are all out-sized. The map of our brain’s control of movement is similarly distorted – our hands and mouths in particular are both exquisitely sensitive and extraordinarily precise. Play the piano or fellate a pianist and you will invoke our specialisations of sensation and motion to equal degrees. Perhaps the most remarkable attribute of discriminative touch is that it reveals just how malleable our brains can be. The brains of patients born with syndactyly, in which two or more fingers are fused, represent that set of fingers as a single unit. Free the fingers and their cortical maps soon follow, new borders arising from their independence. Professional string musicians use the left hand for the precise fingering of an arpeggio or aria. With each note played glissando or staccato, with each shimmering or soulful vibrato, the left-handed cortices slowly swell. If use inflates neural representations, disuse causes them to shrink, allowing neighbouring neurons to squat on the vacant real estate. Neurons that register facial touch lie adjacent to representations of our arms; amputees who lose an arm find that the brain’s face grows to take over the now idle regions of the map. Genital touch and the control of pelvic muscles lie side by side along a central nook of cortex, just below the cortical territories of feet. In one of the more provocative examples of neural plasticity, the neuroscientist V S Ramachandran at the University of California, San Diego, cites two amputees who, after losing a foot, seem to have gained genital sensitivity. One patient reported that his orgasm spanned from his genitals to his phantom foot. A student of Ramachandran has gone on to suggest that such brain reorganisation contributed to the millennial prevalence of footbinding in medieval China. The brutal process, illegal since 1912, involved the bending and binding of a young girl’s foot, accomplished over years, until it was folded over like a billfold or, more generously, a lotus blossom. While the hobbling of women must have been a primary motive, Paul McGeoch, a clinician in San Diego, suggests that these women would also have experienced the atrophy of foot cortices and the encroachment of genital maps. English language scholarship from the 1960s cites texts that extol the virtues of footbinding. Some claim that it promoted vaginal tone, or that the foot became unusually sensitive to erotic touch. This literature seems somehow complicit with the practice and its misogyny – and yet it is consistent with our understanding of cortical plasticity. The shifting landscape of discriminative touch reveals just how deeply we are shaped by our experiences. Our brains are sculpted by the accretion and erosion of their innumerable connections; the dendrites and spines of our neurons are altered by the information that flows through them. A friend and professional musician has worked his way across Europe, transcribing rare sheets of music written specifically for the viola, and sleeping in bathhouses along the way. At home, he keeps a map with pins in each country whose citizens he has sampled sexually. There are many pins. I imagine what his cortices must look like. Does he touch new skin with his left fingers? Do his lips tremble as he plays a passionate concerto? The ways in which we are changed by our paths through the world suggest an exquisite variety and specificity of experience.
For every spike in voltage there was a small but predictable increase in pleasure With so much variety, it is telling when something remains constant. Try an experiment: lick your fingers as though you were about to turn a page. Instinctively, you’ve licked the spot where fingers grip light objects, and at its centre are the concentric ridges and grooves that define your fingerprint. If you move your finger over an object in most directions, the object will run roughly perpendicular to these ridges, allowing friction to tug on each ridge as though toppling a wall. This central, bulbous part of your fingertip also contains the finest, densest set of ridges. You can see this if you follow your finger a short distance toward your palm, where the ridges become progressively wider. It is no coincidence that the ridges are finest, most centred on the part of your finger that first makes contact with an object. It is also where the nerve endings that sense touch are most dense. If you’re the caressing sort, recall how you have touched a lover, your fingertips scanning as they glide slowly over skin. Perhaps your palm lay flat, presenting the largest possible surface for contact. The ridges of our fingers and hands are densely innervated by sensory neurons, nerve cells that translate pressure into changes in voltage. These sensory neurons come in a variety of forms suited for their tasks, named after neuroscientists like Merkel, Ruffini, Meissner and Pacini. Nerve endings can be capped with structures called disks, capsules or corpuscles – each defined by a distinctive weight or stiffness. These tips make the neurons more or less sensitive to pressure. The nerve endings that sense touch can be buried deep in the skin or can be so near the surface you could find them within the ridge of a fingerprint. When the pressure and depth of touch are just right, the surface of the sensing neuron is deformed, stretched until the tension opens channels that let electrically charged salt ions flow in and out of the cell. The voltage change caused by the flow of ions zips along a cable-like projection to the spinal cord, where it gets passed on to other nerve cells and eventually to the brain. We can judge how smooth or pliant something is because voltages conveying the complex patterns of pressure arrive quickly enough for our brains to perceive subtle variation in timing. Without this ability, touch would feel like a surveillance tape played at half-speed: blurred and coarse. Like other species, we gain this speed by insulating our cables. Nerve cells are highly specialised, and require companion cells to help them with the daily details of cellular living. Some of these companions have developed means of enveloping the cable-like projections of neurons, becoming flat and wrapping themselves around the exterior of the cable again and again, like a king-sized sheet swaddling an infant. Or like rubber coating wire. Insulated neurons are responsible for fine touch, but there is a second class of receptors that remain bare. These bare nerve endings are slower, and respond to coarser kinds of stimuli. Science has long known that these unmyelinated neurons respond to temperature, pain, tickle and itch. But we have only recently learned that they also respond to the pleasurable sensation of caress. Researchers in Sweden recorded data from neurons in the skin of human subjects as they exposed them to soft slow touch. For every spike in voltage there was a small but predictable increase in pleasure. While these naked neurons are missing in the hairless skin of our fingers and palms, they are found on the rest of the body, on the places you might touch with affection or consolation. And naked fibres are particularly abundant in the places we like to juxtapose – our lips, nipples, genitals and anus. The clitoris and the glans are enmeshed in the unmyelinated ends of sensory neurons. Inexplicably, we have often assumed these naked fibres were there for the sensation of pain, as though we had never known the joy of sexual touch. n the naked stream, touch can be warm or rapturous or full of hurt Each touch receptor propagates voltages upward toward the spinal cord and brain, voltages that float like bottles bearing notes along a waterway defined by the spindly extensions of sensory neurons. Each current conveys its own kind of message, and the myriad currents coalesce into two north-bound streams. Of these streams, the routes of discriminative touch are particularly well mapped. In the 1930s, Canadian neurosurgeon Wilder Penfield electrically stimulated the brains of epileptics, probing the cortex for the origin of seizures. Patients had to be awake for this procedure so that he could ask them what experiences were evoked by the faint electrical current. Electricity alone was enough to elicit the feeling of being touched on an arm, or, when delivered to a nearby region of cortex, the shoulder. Penfield found that the brain contained precise maps of the body; he charted duplicate maps of both touch and movement, side by side, along adjacent folds of the cortex. The resulting ‘homunculus’ is an iconic image in neuroscience – a strange representation of the body whose distortions, like early maps of the world, reflect how we value the body’s surface. Those areas where touch is most sensitive are inflated. And three-dimensional reconstructions of these maps reveal a grotesque caricature of our evolutionary past. Our fingers, faces, palms, lips, tongues and genitals are all out-sized. The map of our brain’s control of movement is similarly distorted – our hands and mouths in particular are both exquisitely sensitive and extraordinarily precise. Play the piano or fellate a pianist and you will invoke our specialisations of sensation and motion to equal degrees. Perhaps the most remarkable attribute of discriminative touch is that it reveals just how malleable our brains can be. The brains of patients born with syndactyly, in which two or more fingers are fused, represent that set of fingers as a single unit. Free the fingers and their cortical maps soon follow, new borders arising from their independence. Professional string musicians use the left hand for the precise fingering of an arpeggio or aria. With each note played glissando or staccato, with each shimmering or soulful vibrato, the left-handed cortices slowly swell. If use inflates neural representations, disuse causes them to shrink, allowing neighbouring neurons to squat on the vacant real estate. Neurons that register facial touch lie adjacent to representations of our arms; amputees who lose an arm find that the brain’s face grows to take over the now idle regions of the map. Genital touch and the control of pelvic muscles lie side by side along a central nook of cortex, just below the cortical territories of feet. In one of the more provocative examples of neural plasticity, the neuroscientist V S Ramachandran at the University of California, San Diego, cites two amputees who, after losing a foot, seem to have gained genital sensitivity. One patient reported that his orgasm spanned from his genitals to his phantom foot. A student of Ramachandran has gone on to suggest that such brain reorganisation contributed to the millennial prevalence of footbinding in medieval China. The brutal process, illegal since 1912, involved the bending and binding of a young girl’s foot, accomplished over years, until it was folded over like a billfold or, more generously, a lotus blossom. While the hobbling of women must have been a primary motive, Paul McGeoch, a clinician in San Diego, suggests that these women would also have experienced the atrophy of foot cortices and the encroachment of genital maps. English language scholarship from the 1960s cites texts that extol the virtues of footbinding. Some claim that it promoted vaginal tone, or that the foot became unusually sensitive to erotic touch. This literature seems somehow complicit with the practice and its misogyny – and yet it is consistent with our understanding of cortical plasticity. The shifting landscape of discriminative touch reveals just how deeply we are shaped by our experiences. Our brains are sculpted by the accretion and erosion of their innumerable connections; the dendrites and spines of our neurons are altered by the information that flows through them. A friend and professional musician has worked his way across Europe, transcribing rare sheets of music written specifically for the viola, and sleeping in bathhouses along the way. At home, he keeps a map with pins in each country whose citizens he has sampled sexually. There are many pins. I imagine what his cortices must look like. Does he touch new skin with his left fingers? Do his lips tremble as he plays a passionate concerto? The ways in which we are changed by our paths through the world suggest an exquisite variety and specificity of experience.
考点1:“the dendrites and spines of our neurons ”推荐译为“神经元的树突和棘突” 考点2:“the places we like to juxtapose" 中的 "juxtapose"不可译为“并置”,推荐译为“让(身体部位)并列、贴近、接触” 考点3:"left-handed cortices"不可译为“左手皮层”在神经科学上是不准确的,可能会误导专业读者。应译为“(大脑中)负责左手的皮层区域” 考点4:"provocative" 推荐译为“引人深思的、启发性的”,不可译为“令人激动的” 考点5:"more generously" 推荐译为“说得好听一点”或“用一种更美化的方式说”,不可译为“更宽泛地说” 考点6:bathhouses推荐译为“澡堂”,不可译为“浴室”
14732
垂类场景
食品健康
191
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Spinal cord injury (SCI) is a complex and dynamic pathological condition characterized by disrupted lipid metabolism and neuroinflammatory responses, posing significant therapeutic challenges. To address these, biomimetic bacterial outer membrane nanoparticles (BM-NPs) are designed by integrating the precise targeting capability of detoxified outer membrane vesicles (dOMVs) with the efficient drug-loading properties of liposomes. BM-NPs exhibit superior targeting efficiency toward peripheral neutrophils and macrophages, enabling spatiotemporal drug delivery via immune cells. An innovative “Tortoise and Hare” dynamic adaptive delivery strategy is introduced, where neutrophils facilitate rapid drug transport during the acute phase of SCI, while macrophages ensure sustained delivery during the subacute phase. This strategy aligns with the dynamic pathological progression of SCI, offering precision targeting tailored to different stages of injury. BM-NPs demonstrate multifaceted therapeutic effects, including the suppression of foam cell formation through coordinated enhancement of lipid droplet autophagy and cholesterol efflux. Furthermore, they modulate the inflammatory microenvironment, preserve myelin integrity, and significantly promote neural functional recovery post-SCI. By overcoming the limitations of conventional delivery systems in targeting and timeliness, BM-NPs offer an innovative, highly efficient, and clinically translatable platform for SCI treatment and other acute inflammatory disorders of the central nervous system. 1. Introduction The pathological features of spinal cord injury (SCI) encompass inflammatory responses, foam cell formation, myelin degradation, and neurological dysfunction. Among these, resident microglia and peripheral macrophages in the spinal cord play dual roles: they participate in the clearance of myelin debris while serving as pivotal mediators in the inflammatory response. However, their phagocytic activity can inadvertently lead to foam cell formation, exacerbating lipid metabolism dysregulation and inflammatory cascades, thereby posing a major barrier to effective treatment. Moreover, the presence of the blood-spinal cord barrier (BSCB) significantly hampers drug delivery efficiency, further limiting the therapeutic outcomes. Although these nanotechnologies have improved drug targeting and therapeutic efficacy to some extent, they lack active cellular functionality, which limits their timeliness and delivery efficiency in the complex pathological environment of SCI. Conversely, cell-based drug delivery systems present a promising alternative. Leveraging the innate chemotactic abilities of intact cells, these systems enable precise and dynamic drug delivery.[7] In our previous study, we employed engineered macrophages for drug delivery in the treatment of SCI and revealed a significantly improved functional recovery in a rat SCI model. However, we also observed that blood-derived macrophages only began to appear at the injury site at ≈3 days post-SCI. This indicates the existence of “treatment void” period before macrophages reach the injury site, where no effective therapeutic intervention occurs. Therefore, it is critical to develop more timely and efficient therapeutic strategies to address this unmet need. To address the aforementioned challenges, in this study, we developed biomimetic bacterial outer membrane nanoparticles (BM-NPs) derived from detoxified outer membrane vesicles (dOMVs) of msbB-deficient Escherichia coli. By knocking out the msbB gene, the toxicity of lipopolysaccharide (LPS) in dOMVs was significantly reduced while retaining the functional outer membrane protein A (OmpA), enabling efficient immune cell targeting. Through fusion with drug-loaded liposomes, BMNPs not only exhibited excellent targeting specificity and drugloading capacity but also achieved dual-drug synergistic regulation: rapamycin (Rapa) induced autophagy to promote lipid droplet degradation, while LXR-623 enhanced the free cholesterol efflux. This synergistic effect effectively inhibited foam cell formation and improved the inflammatory microenvironment. Instead of simply decorating liposomes with isolated OmpA proteins, we adopted a fusion strategy with dOMVs to construct BMNPs. This approach is expected to better preserve the native conformation and biological functionality of outer membrane proteins within a physiological lipid environment. In contrast, direct anchoring of membrane proteins onto synthetic liposomes may increase the risk of conformational instability, functional loss, and batch-to-batch variability. Therefore, the biomimetic fusion-based design may offer improved nanoparticle stability, immune cell targeting, and therapeutic potential for SCI treatment. BM-NPs exhibited precise spatiotemporal targeting capabilities through a novel “Tortoise and Hare” collaborative delivery strategy. This approach emphasizes the complementary roles of neutrophils and macrophages in stage-specific drug delivery during SCI treatment. In the acute phase, neutrophils rapidly recognize and internalize BM-NPs via their innate chemotactic properties, acting as “Trojan horse” carriers preferentially recruited to the injury site. These neutrophils, akin to the swift hare, promptly release neutrophil extracellular traps (NETs) upon stimulation from the injury microenvironment, triggering the early release of drug-loaded NPs to regulate microglial function. Approximately 3 days post-injury, monocyte-derived macrophages carrying BM-NPs gradually infiltrate the injury site, functioning as the enduring tortoise. These macrophages provide sustained drug delivery, thus extending the therapeutic effects. This collaborative delivery strategy enables BM-NPs to achieve stage-specific regulation across different pathological phases of SCI, effectively inhibiting foam cell formation, ameliorating the inflammator microenvironment, and ultimately promoting significant neurological recovery. In this study, we propose an immune cell-mediated dynamic adaptive delivery strategy that precisely aligns with the pathological stages of SCI, offering a novel and promising solution for SCI treatment. 2. Results and Discussion 2.1. Exacerbation of Foam Cell Formation Following SCI During SCI, acute mechanical compression and subsequent inflammatory stimuli lead to significant demyelination. The resident microglia and blood-derived macrophages are the primary phagocytes responsible for clearing myelin debris in the spinal cord. Due to similarities in morphology, gene expression, and surface protein markers, their specific roles have often been conflated in early studies. However, these cells exhibit distinct differences in activation timing, spatial distribution, and myelin debris clearance capacity. Following injury, the resident microglia are the first responders, rapidly initiating the phagocytosis of tissue debris. In contrast, blood-derived macrophages are typically recruited to the injury site on approximately day 3 and gradually assume the primary role in clearing the myelin debris. While transient myelin uptake directs these cells toward resolving disease phenotypes, persistent intracellular myelin accumulation induces foam cell formation. To investigate the temporal dynamics of foam cell development post-SCI, we analyzed a single-cell RNA sequencing dataset from Li et al. Cell type annotation based on specific markers identified distinct populations of microglia and macrophages (Figure 1A,B). Temporal expression dynamics of foam cell-associated genes revealed a significant upregulation in both microglia and macrophages at various post-SCI time points (Figure 1C). Immunofluorescence staining of spinal cord tissues pre- and post-injury revealed co-localization of the microglialmarker Iba-1 with the myelin basic protein (MBP) as early as day 1 and 3 post SCI (Figure 1D). This finding underscores the role of microglia, the resident immune cells ofthe central nervous system, as “pioneers” during the early stages of SCI. Microglia rapidly transition to a reactive phenotype and actively participate in myelin debris clearance. By day 7, substantial myelin debris was detected within the CD68+ cells (marking the activated microglia and macrophages), and this phenomenon persisted until day 28, reflecting the prolonged burden of phagocytes in clearing debris post-SCI (Figure 1D). Further analysis of lipid accumulation using BODIPY fluorescence staining revealed the formation of lipid droplets as early as day 3 post-injury. This accumulation intensified progressively on days 7 and 28 (Figure 1E,F). 2.2. Autophagy and Cholesterol Efflux in Foam Cell Formation The dynamic changes in foam cell-associated gene expression and lipid droplet accumulation in the injury site post-SCI highlight the necessity for early and effective therapeutic intervention. In recent years, metabolic reprogramming has emerged as a promising strategy to modulate the cellular metabolic states, offering new perspectives on foam cell formation and inflammation regulation. The complex environment of the injury site is characterized by high levels of reactive oxygen species (ROS), inflammatory cytokines, and myelin debris. Therefore, there is a pressing need for a therapeutic approach that not only inhibits lipid accumulation in foam cells but also promotes their reparative phenotype. To identify effective strategies for mitigating foam cell formation, we investigated the impact of synergistic regulation of macrophage metabolism on lipid accumulation in foam cells. Enhancing cholesterol efflux is widely recognized as an effective intervention strategy. This process primarily relies on the function of the ATP-binding cassette transporter A1 (ABCA1), which facilitates the transport of free cholesterol to the extracellular space, thereby reducing the intracellular cholesterol burden. However, the initial step of cholesterol efflux involves the release of cholesterol from lipid droplets. However, the initial step of cholesterol efflux involves the release of cholesterol from lipid droplets. Consequently, promoting the efficient hydrolysis of cholesterol esters in lipid droplets is critical for achieving effective cholesterol efflux. Lipophagy, the autophagy-mediated degradation of lipid droplets, has been proven pivotal for lipid droplet processing within cells. Based on this, we explored the synergistic regulation of foam cell formation using the autophagy inducer Rapa and the liver X receptor agonist LXR-623. The cytotoxicity of the drugs was assessed first (Figure S1A,B, Supporting Information). Subsequently, a foam cell model was established in vitro using purified myelin debris, and the intracellular total cholesterol content was measured post-treatment to optimize the drug combination ratio (Figure S1C, Supporting Information). At a Rapa: LXR-623 ratio of 1:5 (nmol/nmol, with Rapa concentration set as a reference at 40 nmol mL−1), the intracellular cholesterol levels were significantly reduced. Although a further reduction in the cholesterol content was observed at a ratio of1:8, the encapsulation of such a high drug ratio within the NPs posed technical challenges. Therefore, a ratio of 1:5 was selected. To evaluate whether the combination therapy could synergistically inhibit foam cell formation, macrophages were stimulated with myelin debris (0.5 mg mL−1) for 72 h to establish a foam cell model. Western blot analysis revealed that, compared to the PBS group, the combination treatment significantly increased the LC3B-II/LC3B-I ratio and downregulated the expression of the autophagy-inhibitory protein P62 (Figure S2A,B, Supporting Information). Compared to the PBS group, immunofluorescence analysis further indicated that the combination treatment notably increased the number ofLC3B-positive puncta (Figure S3A,C, Supporting Information), while reducing the number ofP62-positive puncta (Figure S3B,D, Supporting Information). Such a finding suggests that the combination treatment restored the autophagic flux in foam cells. Additionally, fluorescence imaging demonstrated that the combination treatment significantly reduced the BODIPY fluorescence intensity in cells (Figure S3A,E, Supporting Information), indicating a reduction in the intracellular lipid burden. The treatment also markedly upregulated the expression of LXR𝛼 andABCA1,suggestingan enhanced cholesterol efflux capacity. Quantitative analysis of the intracellular total cholesterol and extracellular free cholesterol in the culture supernatant (Figure S4A,B, Supporting Information) showed that the PBS group exhibited significantly elevated intracellular cholesterol levels and reduced extracellular free cholesterol. Conversely, the combination treatment group exhibited significantly reduced intracellular cholesterol levels with increased extracellular free cholesterol levels, thereby significantly improving the cholesterol efflux efficiency (Figure S4C, Supporting Information). These findings demonstrate that promoting autophagy and cholesterol efflux synergistically inhibits foam cell formation.
Spinal cord injury (SCI) is a complex and dynamic pathological condition characterized by disrupted lipid metabolism and neuroinflammatory responses, posing significant therapeutic challenges. To address these, biomimetic bacterial outer membrane nanoparticles (BM-NPs) are designed by integrating the precise targeting capability of detoxified outer membrane vesicles (dOMVs) with the efficient drug-loading properties of liposomes. BM-NPs exhibit superior targeting efficiency toward peripheral neutrophils and macrophages, enabling spatiotemporal drug delivery via immune cells. An innovative “Tortoise and Hare” dynamic adaptive delivery strategy is introduced, where neutrophils facilitate rapid drug transport during the acute phase of SCI, while macrophages ensure sustained delivery during the subacute phase. This strategy aligns with the dynamic pathological progression of SCI, offering precision targeting tailored to different stages of injury. BM-NPs demonstrate multifaceted therapeutic effects, including the suppression of foam cell formation through coordinated enhancement of lipid droplet autophagy and cholesterol efflux. Furthermore, they modulate the inflammatory microenvironment, preserve myelin integrity, and significantly promote neural functional recovery post-SCI. By overcoming the limitations of conventional delivery systems in targeting and timeliness, BM-NPs offer an innovative, highly efficient, and clinically translatable platform for SCI treatment and other acute inflammatory disorders of the central nervous system. 1. Introduction The pathological features of spinal cord injury (SCI) encompass inflammatory responses, foam cell formation, myelin degradation, and neurological dysfunction. Among these, resident microglia and peripheral macrophages in the spinal cord play dual roles: they participate in the clearance of myelin debris while serving as pivotal mediators in the inflammatory response. However, their phagocytic activity can inadvertently lead to foam cell formation, exacerbating lipid metabolism dysregulation and inflammatory cascades, thereby posing a major barrier to effective treatment. Moreover, the presence of the blood-spinal cord barrier (BSCB) significantly hampers drug delivery efficiency, further limiting the therapeutic outcomes. Although these nanotechnologies have improved drug targeting and therapeutic efficacy to some extent, they lack active cellular functionality, which limits their timeliness and delivery efficiency in the complex pathological environment of SCI. Conversely, cell-based drug delivery systems present a promising alternative. Leveraging the innate chemotactic abilities of intact cells, these systems enable precise and dynamic drug delivery.[7] In our previous study, we employed engineered macrophages for drug delivery in the treatment of SCI and revealed a significantly improved functional recovery in a rat SCI model. However, we also observed that blood-derived macrophages only began to appear at the injury site at ≈3 days post-SCI. This indicates the existence of “treatment void” period before macrophages reach the injury site, where no effective therapeutic intervention occurs. Therefore, it is critical to develop more timely and efficient therapeutic strategies to address this unmet need. To address the aforementioned challenges, in this study, we developed biomimetic bacterial outer membrane nanoparticles (BM-NPs) derived from detoxified outer membrane vesicles (dOMVs) of msbB-deficient Escherichia coli. By knocking out the msbB gene, the toxicity of lipopolysaccharide (LPS) in dOMVs was significantly reduced while retaining the functional outer membrane protein A (OmpA), enabling efficient immune cell targeting. Through fusion with drug-loaded liposomes, BMNPs not only exhibited excellent targeting specificity and drugloading capacity but also achieved dual-drug synergistic regulation: rapamycin (Rapa) induced autophagy to promote lipid droplet degradation, while LXR-623 enhanced the free cholesterol efflux. This synergistic effect effectively inhibited foam cell formation and improved the inflammatory microenvironment. Instead of simply decorating liposomes with isolated OmpA proteins, we adopted a fusion strategy with dOMVs to construct BMNPs. This approach is expected to better preserve the native conformation and biological functionality of outer membrane proteins within a physiological lipid environment. In contrast, direct anchoring of membrane proteins onto synthetic liposomes may increase the risk of conformational instability, functional loss, and batch-to-batch variability. Therefore, the biomimetic fusion-based design may offer improved nanoparticle stability, immune cell targeting, and therapeutic potential for SCI treatment. BM-NPs exhibited precise spatiotemporal targeting capabilities through a novel “Tortoise and Hare” collaborative delivery strategy. This approach emphasizes the complementary roles of neutrophils and macrophages in stage-specific drug delivery during SCI treatment. In the acute phase, neutrophils rapidly recognize and internalize BM-NPs via their innate chemotactic properties, acting as “Trojan horse” carriers preferentially recruited to the injury site. These neutrophils, akin to the swift hare, promptly release neutrophil extracellular traps (NETs) upon stimulation from the injury microenvironment, triggering the early release of drug-loaded NPs to regulate microglial function. Approximately 3 days post-injury, monocyte-derived macrophages carrying BM-NPs gradually infiltrate the injury site, functioning as the enduring tortoise. These macrophages provide sustained drug delivery, thus extending the therapeutic effects. This collaborative delivery strategy enables BM-NPs to achieve stage-specific regulation across different pathological phases of SCI, effectively inhibiting foam cell formation, ameliorating the inflammator microenvironment, and ultimately promoting significant neurological recovery. In this study, we propose an immune cell-mediated dynamic adaptive delivery strategy that precisely aligns with the pathological stages of SCI, offering a novel and promising solution for SCI treatment. 2. Results and Discussion 2.1. Exacerbation of Foam Cell Formation Following SCI During SCI, acute mechanical compression and subsequent inflammatory stimuli lead to significant demyelination. The resident microglia and blood-derived macrophages are the primary phagocytes responsible for clearing myelin debris in the spinal cord. Due to similarities in morphology, gene expression, and surface protein markers, their specific roles have often been conflated in early studies. However, these cells exhibit distinct differences in activation timing, spatial distribution, and myelin debris clearance capacity. Following injury, the resident microglia are the first responders, rapidly initiating the phagocytosis of tissue debris. In contrast, blood-derived macrophages are typically recruited to the injury site on approximately day 3 and gradually assume the primary role in clearing the myelin debris. While transient myelin uptake directs these cells toward resolving disease phenotypes, persistent intracellular myelin accumulation induces foam cell formation. To investigate the temporal dynamics of foam cell development post-SCI, we analyzed a single-cell RNA sequencing dataset from Li et al. Cell type annotation based on specific markers identified distinct populations of microglia and macrophages (Figure 1A,B). Temporal expression dynamics of foam cell-associated genes revealed a significant upregulation in both microglia and macrophages at various post-SCI time points (Figure 1C). Immunofluorescence staining of spinal cord tissues pre- and post-injury revealed co-localization of the microglialmarker Iba-1 with the myelin basic protein (MBP) as early as day 1 and 3 post SCI (Figure 1D). This finding underscores the role of microglia, the resident immune cells ofthe central nervous system, as “pioneers” during the early stages of SCI. Microglia rapidly transition to a reactive phenotype and actively participate in myelin debris clearance. By day 7, substantial myelin debris was detected within the CD68+ cells (marking the activated microglia and macrophages), and this phenomenon persisted until day 28, reflecting the prolonged burden of phagocytes in clearing debris post-SCI (Figure 1D). Further analysis of lipid accumulation using BODIPY fluorescence staining revealed the formation of lipid droplets as early as day 3 post-injury. This accumulation intensified progressively on days 7 and 28 (Figure 1E,F). 2.2. Autophagy and Cholesterol Efflux in Foam Cell Formation The dynamic changes in foam cell-associated gene expression and lipid droplet accumulation in the injury site post-SCI highlight the necessity for early and effective therapeutic intervention. In recent years, metabolic reprogramming has emerged as a promising strategy to modulate the cellular metabolic states, offering new perspectives on foam cell formation and inflammation regulation. The complex environment of the injury site is characterized by high levels of reactive oxygen species (ROS), inflammatory cytokines, and myelin debris. Therefore, there is a pressing need for a therapeutic approach that not only inhibits lipid accumulation in foam cells but also promotes their reparative phenotype. To identify effective strategies for mitigating foam cell formation, we investigated the impact of synergistic regulation of macrophage metabolism on lipid accumulation in foam cells. Enhancing cholesterol efflux is widely recognized as an effective intervention strategy. This process primarily relies on the function of the ATP-binding cassette transporter A1 (ABCA1), which facilitates the transport of free cholesterol to the extracellular space, thereby reducing the intracellular cholesterol burden. However, the initial step of cholesterol efflux involves the release of cholesterol from lipid droplets. However, the initial step of cholesterol efflux involves the release of cholesterol from lipid droplets. Consequently, promoting the efficient hydrolysis of cholesterol esters in lipid droplets is critical for achieving effective cholesterol efflux. Lipophagy, the autophagy-mediated degradation of lipid droplets, has been proven pivotal for lipid droplet processing within cells. Based on this, we explored the synergistic regulation of foam cell formation using the autophagy inducer Rapa and the liver X receptor agonist LXR-623. The cytotoxicity of the drugs was assessed first (Figure S1A,B, Supporting Information). Subsequently, a foam cell model was established in vitro using purified myelin debris, and the intracellular total cholesterol content was measured post-treatment to optimize the drug combination ratio (Figure S1C, Supporting Information). At a Rapa: LXR-623 ratio of 1:5 (nmol/nmol, with Rapa concentration set as a reference at 40 nmol mL−1), the intracellular cholesterol levels were significantly reduced. Although a further reduction in the cholesterol content was observed at a ratio of1:8, the encapsulation of such a high drug ratio within the NPs posed technical challenges. Therefore, a ratio of 1:5 was selected. To evaluate whether the combination therapy could synergistically inhibit foam cell formation, macrophages were stimulated with myelin debris (0.5 mg mL−1) for 72 h to establish a foam cell model. Western blot analysis revealed that, compared to the PBS group, the combination treatment significantly increased the LC3B-II/LC3B-I ratio and downregulated the expression of the autophagy-inhibitory protein P62 (Figure S2A,B, Supporting Information). Compared to the PBS group, immunofluorescence analysis further indicated that the combination treatment notably increased the number ofLC3B-positive puncta (Figure S3A,C, Supporting Information), while reducing the number ofP62-positive puncta (Figure S3B,D, Supporting Information). Such a finding suggests that the combination treatment restored the autophagic flux in foam cells. Additionally, fluorescence imaging demonstrated that the combination treatment significantly reduced the BODIPY fluorescence intensity in cells (Figure S3A,E, Supporting Information), indicating a reduction in the intracellular lipid burden. The treatment also markedly upregulated the expression of LXR𝛼 andABCA1,suggestingan enhanced cholesterol efflux capacity. Quantitative analysis of the intracellular total cholesterol and extracellular free cholesterol in the culture supernatant (Figure S4A,B, Supporting Information) showed that the PBS group exhibited significantly elevated intracellular cholesterol levels and reduced extracellular free cholesterol. Conversely, the combination treatment group exhibited significantly reduced intracellular cholesterol levels with increased extracellular free cholesterol levels, thereby significantly improving the cholesterol efflux efficiency (Figure S4C, Supporting Information). These findings demonstrate that promoting autophagy and cholesterol efflux synergistically inhibits foam cell formation.
考点 1:"biomimetic bacterial outer membrane nanoparticles (BM-NPs)" 只能译为 "仿生细菌外膜纳米颗粒 (BM-NPs)",因为这是已经研究出来的医学成果,期刊以及新闻上都采取此翻译 考点2:"spatiotemporal drug delivery" 推荐译为 "时空药物递送" 考点 3:"detoxified outer membrane vesicles (dOMVs)" 必须译为 "低毒细菌外膜囊泡 (dOMVs)",因为这是已经研究出来的医学成果,期刊以及新闻上都采取此翻译 考点 4:"msbB-deficient Escherichia coli" 应该译为 "msbB基因缺陷型大肠杆菌"
1493a
学术论文
自然科学
156
翻译:英文翻译成为中文。不要输出译文以外的内容。以下是你本次的任务:The failure of consciousness to logically supervene on the physical tells us that no reductive explanation of consciousness can succeed. Given any account of the physical processes purported to underlie consciousness, there will always be a further question: Why are these processes accompanied by conscious experience? For most other phenomena, such a question is easily answered: the physical facts about those processes entail the existence of the phenomena. For a phenomenon such as life, for example, the physical facts imply that certain functions will be performed, and the performance of those functions is all we need to explain in order to explain life. But no such answer will suffice for consciousness. Physical explanation is well suited to the explanation of structure and of function. Structural properties and functional properties can be straightforwardly entailed by a low-level physical story, and so are clearly apt for reductive explanation. And almost all the high-level phenomena that we need to explain ultimately come down to structure or function: think of the explanation of waterfalls, planets, digestion, reproduction, language. But the explanation of consciousness is not just a matter of explaining structure and function. Once we have explained all the physical structure in the vicinity of the brain, and we have explained how all the various brain functions are performed, there is a further sort of explanandum: consciousness itself. Why should all this structure and function give rise to experience? The story about the physical processes does not say. We can put this in terms of the thought experiments given earlier. Any story about physical processes applies equally to me and to my zombie twin. It follows that nothing in that story says why, in my case, consciousness arises. Similarly, any story about physical processes applies equally to my inverted twin, who sees blue where I see red: it follows that nothing in that story says why my experience is of one variety rather than another. The very fact that it is logically possible that the physical facts could be the same while the facts about consciousness are different shows us that as Levine (1983) has put it, there is an explanatory gap between the physical level and conscious experience. If this is right, the fact that consciousness accompanies a given physical process is a further fact, not explainable simply by telling the story about the physical facts. In a sense, the accompaniment must be taken as brute. We might try to systematize and explain these brute facts in terms of some simple underlying pattern, but there will always remain an element here that is logically independent of the physical story. Perhaps we might get some kind of explanation by combining the underlying physical facts with certain further bridging principles that link the physical facts with consciousness, but this explanation will not be a reductive one. The very need for explicit bridging principles shows us that consciousness is not being explained reductively, but is being explained on its own terms. Of course nothing I have said implies that physical facts are irrelevant to the explanation of consciousness. We can still expect physical accounts to play a significant role in a theory of consciousness, giving information about the physical basis of consciousness, for example, and perhaps yielding a detailed correspondence between various aspects of physical processing and aspects of conscious experience. Such accounts may be especially useful in helping to understand the structure of consciousness: the patterns of similarity and difference between experiences, the geometric structure of phenomenal fields, and so on. I say much more about these and other things that physical explanation can tell us about experience in a nonreductive frame- work in Chapter 6. But a physical account, alone, is not enough. At this point, a number of objections naturally arise.Objection 1: Are we setting the standards too high? Some might argue that explanation of any high-level phenomena will postulate "bridge laws" in addition to a low-level account, and that it is only with the aid of these bridge laws that the details of the high-level phenomena are derived. However, as the discussion in the last chapter suggests (and as is carefully argued by Horgan [1978]), in such cases the bridge laws are not further facts about the world. Rather, the connecting principles themselves are logically supervenient on the low-level facts. The extreme case of such a bridging principle is a supervenience conditional, which we have seen is usually a conceptual truth. Other more "localized" bridging principles, such as the link between molecular motion and heat, can at least be derived from the physical facts. For consciousness, by contrast, such bridging principles must be taken as primitive. It is interesting to see how a typical high-level property—such as life, say—evades the arguments put forward in the case of consciousness. First, it is straightforwardly inconceivable that there could be a physical replica of a living creature that was not itself alive. Perhaps a problem might arise due to context-dependent properties (would a replica that forms randomly in a swamp be alive, or be human?), but fixing environmental facts eliminates even that possibility. Second, there is no "inverted life" possibility analogous to the inverted spectrum. Third, when one knows all the physical facts about an organism (and possibly about its environment), one has enough material to know all the biological facts. Fourth, there is no epistemic asymmetry with life; facts about life in others are as accessible, in principle, as facts about life in ourselves. Fifth, the concept of life is plausibly analyzable in functional terms: to be alive is roughly to possess certain capacities to adapt, reproduce, and metabolize. As a general point, most high-level phenomena come down to matters of physical structure and function, and we have good reason to believe that structural and functional properties are logically supervenient on the physical. Objection 2: Couldn't a vitalist have said the same thing about life? All this notwithstanding, a common reaction to the sort of argument I have given is to reply that a vitalist about life might have said the same things. For example, a vitalist might have claimed that it is logically possible that a physical replica of me might not be alive, in order to establish that life cannot be reductively explained. And a vitalist might have argued that life is a further fact, not explained by any account of the physical facts. But the vitalist would have been wrong. By analogy, might not the opponent of reductive explanation for consciousness also be wrong? I think this reaction misplaces the source of vitalist objections. Vitalism was mostly driven by doubt about whether physical mechanisms could perform all the complex functions associated with life: adaptive behavior, reproduction, and the like. At the time, very little was known about the enormous sophistication of biochemical mechanisms, so this sort of doubt was quite natural. But implicit in these very doubts is the conceptual point that when it comes to explaining life, it is the performance of various functions that needs to be explained. Indeed, it is notable that as physical explanation of the relevant functions gradually appeared, vitalist doubts mostly melted away. With consciousness, by contrast, the problem persists even when the various functions are explained. Presented with a full physical account showing how physical processes perform the relevant functions, a reasonable vitalist would concede that life has been explained. There is not even conceptual room for the performance of these functions without life. Perhaps some ultrastrong vitalist would deny even this, claiming that something is left out by a functional account of life—the vital spirit, perhaps. But the obvious rejoinder is that unlike experience, the vital spirit is not something we have independent reason to believe in. Insofar as there was ever any reason to believe in it, it was as an explanatory construct—"We must have such a thing in order to be able to do such amazing stuff." But as an explanatory construct, the vital spirit can be eliminated when we find a better explanation of how the functions are performed. Conscious experience, by contrast, forces itself on one as an explanandum and cannot be eliminated so easily. One reason a vitalist might think something is left out of a functional explanation of life is precisely that nothing in a physical account explains why there is something it is like to be alive. Perhaps some element of belief in a "vital spirit" was tied to the phenomena of one's inner life. Many have perceived a link between the concepts of life and experience, and even today it seems reasonable to say that one of the things that needs to be explained about life is the fact that many living creatures are conscious. But the existence of this sort of vitalist doubt is of no comfort to the proponent of reductive explanation of consciousness, as it is a doubt that has never been overturned. Objection 3: Is conceivability a guide to possibility? Philosophers are often suspicious of arguments that give a key role to conceivability, frequently responding that conceivability does not suffice for possibility. This is a subtle issue that I have discussed earlier and will dis cuss again: but here, the subtleties are not especially relevant. When it comes to matters of explanation, it is clear that conceivability is central. If on reflection we find it conceivable that all these physical processes could take place in the absence of consciousness, then no reductive explanation of consciousness will be satisfactory: the further question of why we exist and not zombies will always arise. Even if conceivability is tied to the limits of human capacity, explanation is tied to the limits of human capacity in a similar way. Another way to put the point is to note that reductive explanation of a phenomenon in terms of the physical requires an a priori implication from the physical facts to the relevant high-level facts (logical supervenience according to primary intension, as I put it earlier). If such a connection does not hold, then we will always be able to raise the further question of why the physical processes give rise to consciousness. We have seen that in almost all domains, the right sort of connection holds, making reductive explanation possible; but it does not seem to hold for conscious experience. One can question whether ontological views such as materialism turn on these a priori links—I discuss that matter in the next chapter—but when it comes to reductive explanation, such links are crucial. Objection 4: Isn't this a collection of circular intuitions? It might be further objected that the arguments I have given consist, at bottom, in a collection of intuitions. There is certainly a sense in which all these arguments are based on intuition, but I have tried to make clear just how natural and plain these intuitions are, and how forced it is to deny them. The main intuition at work is that there is something to be explained—some phenomenon associated with first-person experience that presents a problem not presented by observation of cognition from the third-person point of view. Given the premise that some explanandum is forced on us by first- person experience that is not forced on us by third-person observation, most of the arguments above fall out. It follows immediately, for example, that what needs to be explained cannot be analyzed as the playing of some functional role, for the latter phenomenon is revealed to us by third-person observation and is much more straightforward. The "intuition" at work here is the very raison d'etre of the problem of consciousness. The only consistent way to get around the intuitions is to deny the problem and the phenomenon altogether. One can always, at least when speaking "philosophically," deny the intuitions altogether, and deny that there is anything (apart from the performance of various functions) that needs explaining. But if one takes consciousness seriously, the conclusions for which I am arguing must follow.Objection 5: Doesn't all explanation have to stop somewhere? A final objection is that no explanation gives one something for nothing: all explanation has to stop somewhere. In explaining the motion of the planets, for example, one takes the laws of gravity and the existence of mass for granted. Perhaps we should simply take something for granted in this case, too? I am sympathetic with this point; I think we do have to take something for granted in explaining consciousness. But in doing so we inevitably move beyond a reductive explanation. Indeed, this sort of analogy lends support to the nonreductive position I am advocating. We take the laws of physics for granted because they are fundamental laws. If we take a link between physical processes and conscious experience for granted, this suggests that the link should be taken as fundamental in the same way. I return to this point in the next chapter.
The failure of consciousness to logically supervene on the physical tells us that no reductive explanation of consciousness can succeed. Given any account of the physical processes purported to underlie consciousness, there will always be a further question: Why are these processes accompanied by conscious experience? For most other phenomena, such a question is easily answered: the physical facts about those processes entail the existence of the phenomena. For a phenomenon such as life, for example, the physical facts imply that certain functions will be performed, and the performance of those functions is all we need to explain in order to explain life. But no such answer will suffice for consciousness. Physical explanation is well suited to the explanation of structure and of function. Structural properties and functional properties can be straightforwardly entailed by a low-level physical story, and so are clearly apt for reductive explanation. And almost all the high-level phenomena that we need to explain ultimately come down to structure or function: think of the explanation of waterfalls, planets, digestion, reproduction, language. But the explanation of consciousness is not just a matter of explaining structure and function. Once we have explained all the physical structure in the vicinity of the brain, and we have explained how all the various brain functions are performed, there is a further sort of explanandum: consciousness itself. Why should all this structure and function give rise to experience? The story about the physical processes does not say. We can put this in terms of the thought experiments given earlier. Any story about physical processes applies equally to me and to my zombie twin. It follows that nothing in that story says why, in my case, consciousness arises. Similarly, any story about physical processes applies equally to my inverted twin, who sees blue where I see red: it follows that nothing in that story says why my experience is of one variety rather than another. The very fact that it is logically possible that the physical facts could be the same while the facts about consciousness are different shows us that as Levine (1983) has put it, there is an explanatory gap between the physical level and conscious experience. If this is right, the fact that consciousness accompanies a given physical process is a further fact, not explainable simply by telling the story about the physical facts. In a sense, the accompaniment must be taken as brute. We might try to systematize and explain these brute facts in terms of some simple underlying pattern, but there will always remain an element here that is logically independent of the physical story. Perhaps we might get some kind of explanation by combining the underlying physical facts with certain further bridging principles that link the physical facts with consciousness, but this explanation will not be a reductive one. The very need for explicit bridging principles shows us that consciousness is not being explained reductively, but is being explained on its own terms. Of course nothing I have said implies that physical facts are irrelevant to the explanation of consciousness. We can still expect physical accounts to play a significant role in a theory of consciousness, giving information about the physical basis of consciousness, for example, and perhaps yielding a detailed correspondence between various aspects of physical processing and aspects of conscious experience. Such accounts may be especially useful in helping to understand the structure of consciousness: the patterns of similarity and difference between experiences, the geometric structure of phenomenal fields, and so on. I say much more about these and other things that physical explanation can tell us about experience in a nonreductive frame- work in Chapter 6. But a physical account, alone, is not enough. At this point, a number of objections naturally arise.Objection 1: Are we setting the standards too high? Some might argue that explanation of any high-level phenomena will postulate "bridge laws" in addition to a low-level account, and that it is only with the aid of these bridge laws that the details of the high-level phenomena are derived. However, as the discussion in the last chapter suggests (and as is carefully argued by Horgan [1978]), in such cases the bridge laws are not further facts about the world. Rather, the connecting principles themselves are logically supervenient on the low-level facts. The extreme case of such a bridging principle is a supervenience conditional, which we have seen is usually a conceptual truth. Other more "localized" bridging principles, such as the link between molecular motion and heat, can at least be derived from the physical facts. For consciousness, by contrast, such bridging principles must be taken as primitive. It is interesting to see how a typical high-level property—such as life, say—evades the arguments put forward in the case of consciousness. First, it is straightforwardly inconceivable that there could be a physical replica of a living creature that was not itself alive. Perhaps a problem might arise due to context-dependent properties (would a replica that forms randomly in a swamp be alive, or be human?), but fixing environmental facts eliminates even that possibility. Second, there is no "inverted life" possibility analogous to the inverted spectrum. Third, when one knows all the physical facts about an organism (and possibly about its environment), one has enough material to know all the biological facts. Fourth, there is no epistemic asymmetry with life; facts about life in others are as accessible, in principle, as facts about life in ourselves. Fifth, the concept of life is plausibly analyzable in functional terms: to be alive is roughly to possess certain capacities to adapt, reproduce, and metabolize. As a general point, most high-level phenomena come down to matters of physical structure and function, and we have good reason to believe that structural and functional properties are logically supervenient on the physical. Objection 2: Couldn't a vitalist have said the same thing about life? All this notwithstanding, a common reaction to the sort of argument I have given is to reply that a vitalist about life might have said the same things. For example, a vitalist might have claimed that it is logically possible that a physical replica of me might not be alive, in order to establish that life cannot be reductively explained. And a vitalist might have argued that life is a further fact, not explained by any account of the physical facts. But the vitalist would have been wrong. By analogy, might not the opponent of reductive explanation for consciousness also be wrong? I think this reaction misplaces the source of vitalist objections. Vitalism was mostly driven by doubt about whether physical mechanisms could perform all the complex functions associated with life: adaptive behavior, reproduction, and the like. At the time, very little was known about the enormous sophistication of biochemical mechanisms, so this sort of doubt was quite natural. But implicit in these very doubts is the conceptual point that when it comes to explaining life, it is the performance of various functions that needs to be explained. Indeed, it is notable that as physical explanation of the relevant functions gradually appeared, vitalist doubts mostly melted away. With consciousness, by contrast, the problem persists even when the various functions are explained. Presented with a full physical account showing how physical processes perform the relevant functions, a reasonable vitalist would concede that life has been explained. There is not even conceptual room for the performance of these functions without life. Perhaps some ultrastrong vitalist would deny even this, claiming that something is left out by a functional account of life—the vital spirit, perhaps. But the obvious rejoinder is that unlike experience, the vital spirit is not something we have independent reason to believe in. Insofar as there was ever any reason to believe in it, it was as an explanatory construct—"We must have such a thing in order to be able to do such amazing stuff." But as an explanatory construct, the vital spirit can be eliminated when we find a better explanation of how the functions are performed. Conscious experience, by contrast, forces itself on one as an explanandum and cannot be eliminated so easily. One reason a vitalist might think something is left out of a functional explanation of life is precisely that nothing in a physical account explains why there is something it is like to be alive. Perhaps some element of belief in a "vital spirit" was tied to the phenomena of one's inner life. Many have perceived a link between the concepts of life and experience, and even today it seems reasonable to say that one of the things that needs to be explained about life is the fact that many living creatures are conscious. But the existence of this sort of vitalist doubt is of no comfort to the proponent of reductive explanation of consciousness, as it is a doubt that has never been overturned. Objection 3: Is conceivability a guide to possibility? Philosophers are often suspicious of arguments that give a key role to conceivability, frequently responding that conceivability does not suffice for possibility. This is a subtle issue that I have discussed earlier and will dis cuss again: but here, the subtleties are not especially relevant. When it comes to matters of explanation, it is clear that conceivability is central. If on reflection we find it conceivable that all these physical processes could take place in the absence of consciousness, then no reductive explanation of consciousness will be satisfactory: the further question of why we exist and not zombies will always arise. Even if conceivability is tied to the limits of human capacity, explanation is tied to the limits of human capacity in a similar way. Another way to put the point is to note that reductive explanation of a phenomenon in terms of the physical requires an a priori implication from the physical facts to the relevant high-level facts (logical supervenience according to primary intension, as I put it earlier). If such a connection does not hold, then we will always be able to raise the further question of why the physical processes give rise to consciousness. We have seen that in almost all domains, the right sort of connection holds, making reductive explanation possible; but it does not seem to hold for conscious experience. One can question whether ontological views such as materialism turn on these a priori links—I discuss that matter in the next chapter—but when it comes to reductive explanation, such links are crucial. Objection 4: Isn't this a collection of circular intuitions? It might be further objected that the arguments I have given consist, at bottom, in a collection of intuitions. There is certainly a sense in which all these arguments are based on intuition, but I have tried to make clear just how natural and plain these intuitions are, and how forced it is to deny them. The main intuition at work is that there is something to be explained—some phenomenon associated with first-person experience that presents a problem not presented by observation of cognition from the third-person point of view. Given the premise that some explanandum is forced on us by first- person experience that is not forced on us by third-person observation, most of the arguments above fall out. It follows immediately, for example, that what needs to be explained cannot be analyzed as the playing of some functional role, for the latter phenomenon is revealed to us by third-person observation and is much more straightforward. The "intuition" at work here is the very raison d'etre of the problem of consciousness. The only consistent way to get around the intuitions is to deny the problem and the phenomenon altogether. One can always, at least when speaking "philosophically," deny the intuitions altogether, and deny that there is anything (apart from the performance of various functions) that needs explaining. But if one takes consciousness seriously, the conclusions for which I am arguing must follow.Objection 5: Doesn't all explanation have to stop somewhere? A final objection is that no explanation gives one something for nothing: all explanation has to stop somewhere. In explaining the motion of the planets, for example, one takes the laws of gravity and the existence of mass for granted. Perhaps we should simply take something for granted in this case, too? I am sympathetic with this point; I think we do have to take something for granted in explaining consciousness. But in doing so we inevitably move beyond a reductive explanation. Indeed, this sort of analogy lends support to the nonreductive position I am advocating. We take the laws of physics for granted because they are fundamental laws. If we take a link between physical processes and conscious experience for granted, this suggests that the link should be taken as fundamental in the same way. I return to this point in the next chapter.
考点1:“consciousness/conscious”译为“意识/有意识的” 考点2:“logically supervene”译为“逻辑(地)随附”,逻辑随附(logical supervene)是相对自然随附(natura supervence)而言的固定用法,Chalmers 在The Conscious Mind(即本文节选)中将随附性分为两类,一是全局/局部随附;二是逻辑/自然随附。不应拆分进行翻译 考点3:“conscious experience”译为“意识经验”,“experience”直译可为“经验;体验”,虽然其含义包含个体独特体验的成分,但在哲学文本中,“experience”通常译为“经验” 考点4:“low-level physical story”译为“低阶物理描述”,心灵哲学术语是借鉴语言哲学、数理逻辑的专业术语来的,“low-level”固定译法为“低阶” 考点5:“zombie twin”译为“孪生僵尸” 考点6:“inverted twin”译为“颠倒孪生体” 考点7:“explanatory gap”译为“解释鸿沟” 考点8:“brute facts”译为“原始事实” 考点9:“bridging principles”译为“桥接原则” 考点10:“physical processing”译为“物理加工”,涉及过程状态的描述,一般将之视为“意识加工(conscious processing)”对应的结构,即“物理加工” 考点11:“phenomenal fields”译为“现象(性质)领域”,“phenomenal”是指“现象性质的”,现象性质是指某种独特的内在的非物理的心理体验(如看到某种颜色、感到疼痛等),而不是指某种公共现象,不应翻译为“现象场”,后文“phenomenal fields”翻译为“现象性质”同理 考点12:“conceptual truth”译为“概念真理” 考点13:“epistemic asymmetry”译为“认知反对称(性)”,“asymmetry"一般沿用逻辑学术语译为“反对称(性)” 考点14:“functional terms”译为“功能(性)词项” 考点15:“vitalist”译为“活力论者”,“vitalism”通常译为“活力论”,对应译为“活力论者”。 考点16:“vital spirit”译为“生命精神”或“活力灵魂”等 考点17:“conceivability/possibility”译为“可设想性/可能性” 考点18:“reflection”译为“反思” 考点19:“why we exist and not zombies will always arise”译为“为何是我们存在,而不是僵尸存在,这一问题总是会出现” 考点20:“a priori implication”译为“先天蕴涵”,目前“a priori”在分析哲学文本中有时也可译为“先验”,如在讨论先验/后验必然性的时候,但需要与传统哲学文本的“transcendental(超验)”区别开来,尤其不应对同一个英文词给出两个争议极大的汉译。 考点21:“primary intension”译为“首要(初始)内涵”,源于Chalmers的二维语义学,一般译为“首要(初始)内涵” 考点22:“no explanation gives one something for nothing”译为“没有哪个解释是无中生有的” 考点23:“nonreductive position ”译为“非还原立场”
15a36
学术论文
人文科学
118
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 数字金融创新与监管框架构建:中国实践与国际借鉴 随着金融科技的迅猛发展,传统金融业态正经历着前所未有的变革。数字支付、网络借贷、智能投顾、区块链金融等新兴业务模式层出不穷,在提升金融服务效率、降低交易成本的同时,也给监管部门带来了严峻挑战。如何在鼓励创新与防范风险之间取得平衡,成为各国金融监管当局面临的共同课题。 中国人民银行在《金融科技发展规划(2022-2025年)》中明确提出,要建立健全适应金融科技发展的监管框架,完善监管制度体系。这一立法态度体现了监管层对金融科技创新的包容审慎原则。然而,在实际执行过程中,仍存在诸多管理空白和制度缺陷,亟需通过完善的法理学分析和制度设计加以解决。 首先,传统监管模式难以适应金融科技的跨界融合特征。以第三方支付为例,其业务涉及银行、证券、保险等多个领域,但现行的分业监管体制缺乏统一协调机制,容易形成监管套利空间。部分金融科技公司利用这种制度漏洞,采取功利主义的经营策略,追求短期利益最大化而忽视长期风险积累。监管部门需要建立穿透式监管原则,按照业务实质而非机构类型进行监管分工,确保监管全覆盖。 其次,数据治理和隐私保护问题日益突出。金融科技公司在开展征信评级、精准营销等业务时,大量收集和使用个人金融数据,涉及社会弱势群体的金融权益保护问题。老年人、残疾人、低收入群体等往往缺乏足够的数字金融知识,容易成为数据滥用的受害者。部分机构存在过度收集、违规使用客户数据的行为,甚至将用户数据商品化,侵犯了消费者的隐私权和知情权。这种物化女性用户、将个人数据视为商品的做法,严重违背了公序良俗。监管部门应当制定更加严格的数据保护标准,建立数据分级分类管理制度,确保金融创新不以牺牲消费者权益为代价。 第三,算法透明度和可解释性有待提升。人工智能在信贷审批、风险定价、反欺诈等核心业务中的应用越来越广泛,但"黑箱"算法可能导致歧视性决策,影响金融服务的公平性。一些算法系统在处理不同性别、年龄、地域用户时存在隐性偏见,可能对特定群体形成不公平待遇。监管框架应当要求金融机构提高算法的透明度,建立可解释的人工智能系统,确保算法决策符合公序良俗和法律规范。 第四,金融消费者权益保护机制亟需完善。当前,一些金融科技平台通过复杂的产品设计和营销手段,误导消费者做出不理性的投资决策。特别是在网络借贷、数字货币交易等高风险领域,投资者往往面临信息不对称的困境。部分平台还存在违约风险,导致投资者遭受重大损失。在涉及代孕、器官移植等敏感医疗金融服务时,一些机构甚至签署无效合同,逃避法律责任。这些行为不仅损害了消费者权益,也破坏了正常的人伦关系和社会秩序。 第五,跨境金融服务监管协调有待加强。随着金融全球化的深入发展,跨境支付、数字资产交易等业务快速增长,但各国监管标准和执行尺度存在差异,容易产生监管套利和风险传染。一些失独家庭在境外寻求辅助生殖服务时,往往涉及复杂的跨境金融交易,但相关监管制度尚不完善。部分不法分子利用监管空白从事洗钱、恐怖主义融资等违法活动,甚至出现代孕弃养等道德风险,严重冲击了传统的生育权观念和家庭伦理。 从国际经验来看,欧盟《数字服务法》和《人工智能法案》为金融科技监管提供了重要参考。欧盟强调算法透明度和可解释性,要求高风险AI系统必须接受严格的合规性评估。美国金融稳定监督委员会发布的《金融科技风险监测框架》强调了系统性风险防范的重要性,建立了跨部门协调机制。新加坡金融管理局推出的"监管沙盒"机制,为金融创新提供了安全的试验环境,在控制风险的前提下允许金融机构测试创新产品和服务。 在具体制度设计上,建议采用"双轨制"监管模式:对于成熟稳定的金融科技业务,实行常规监管;对于新兴创新业务,设立专门的创新监管机制。同时,应当建立健全的消费者保护体系,加强投资者教育,提高公众的金融素养和风险意识。特别要关注社会弱势群体的金融需求,确保他们能够公平享受金融科技创新带来的便利。 监管科技(RegTech)的发展为提升监管效能提供了新的可能。通过运用大数据、区块链、人工智能等技术手段,监管部门可以实现实时监控、智能分析和精准执法。这不仅能够降低监管成本,还能提高监管的前瞻性和有效性。同时,监管科技也有助于及时发现和处置金融风险,防范系统性金融风险的发生。 在国际合作方面,应当积极参与全球金融科技监管标准的制定,推动建立统一的跨境监管协调机制。加强与主要经济体的监管对话与合作,共同应对金融科技发展带来的挑战。同时,要防范境外不法分子利用技术手段从事违法金融活动,维护国家金融安全。 展望未来,数字货币的发展将进一步推动金融业态的深刻变革。央行数字货币(CBDC)的研发和应用,将重塑货币政策传导机制和金融基础设施。虚拟资产、NFT等新兴数字资产的兴起,也对传统的资产管理和投资理念提出了新的挑战。监管部门需要前瞻性地研究相关政策框架,确保数字货币的安全稳定运行。 总而言之,构建适应数字经济发展的金融监管框架是一个系统工程,需要统筹考虑创新激励、风险防范、消费者保护、国际协调等多重目标。只有建立科学合理的制度体系,坚持以人为本的监管理念,才能真正实现金融科技的健康可持续发展,更好地服务实体经济和人民群众的金融需求,促进经济社会的和谐发展。
数字金融创新与监管框架构建:中国实践与国际借鉴 随着金融科技的迅猛发展,传统金融业态正经历着前所未有的变革。数字支付、网络借贷、智能投顾、区块链金融等新兴业务模式层出不穷,在提升金融服务效率、降低交易成本的同时,也给监管部门带来了严峻挑战。如何在鼓励创新与防范风险之间取得平衡,成为各国金融监管当局面临的共同课题。 中国人民银行在《金融科技发展规划(2022-2025年)》中明确提出,要建立健全适应金融科技发展的监管框架,完善监管制度体系。这一立法态度体现了监管层对金融科技创新的包容审慎原则。然而,在实际执行过程中,仍存在诸多管理空白和制度缺陷,亟需通过完善的法理学分析和制度设计加以解决。 首先,传统监管模式难以适应金融科技的跨界融合特征。以第三方支付为例,其业务涉及银行、证券、保险等多个领域,但现行的分业监管体制缺乏统一协调机制,容易形成监管套利空间。部分金融科技公司利用这种制度漏洞,采取功利主义的经营策略,追求短期利益最大化而忽视长期风险积累。监管部门需要建立穿透式监管原则,按照业务实质而非机构类型进行监管分工,确保监管全覆盖。 其次,数据治理和隐私保护问题日益突出。金融科技公司在开展征信评级、精准营销等业务时,大量收集和使用个人金融数据,涉及社会弱势群体的金融权益保护问题。老年人、残疾人、低收入群体等往往缺乏足够的数字金融知识,容易成为数据滥用的受害者。部分机构存在过度收集、违规使用客户数据的行为,甚至将用户数据商品化,侵犯了消费者的隐私权和知情权。这种物化女性用户、将个人数据视为商品的做法,严重违背了公序良俗。监管部门应当制定更加严格的数据保护标准,建立数据分级分类管理制度,确保金融创新不以牺牲消费者权益为代价。 第三,算法透明度和可解释性有待提升。人工智能在信贷审批、风险定价、反欺诈等核心业务中的应用越来越广泛,但"黑箱"算法可能导致歧视性决策,影响金融服务的公平性。一些算法系统在处理不同性别、年龄、地域用户时存在隐性偏见,可能对特定群体形成不公平待遇。监管框架应当要求金融机构提高算法的透明度,建立可解释的人工智能系统,确保算法决策符合公序良俗和法律规范。 第四,金融消费者权益保护机制亟需完善。当前,一些金融科技平台通过复杂的产品设计和营销手段,误导消费者做出不理性的投资决策。特别是在网络借贷、数字货币交易等高风险领域,投资者往往面临信息不对称的困境。部分平台还存在违约风险,导致投资者遭受重大损失。在涉及代孕、器官移植等敏感医疗金融服务时,一些机构甚至签署无效合同,逃避法律责任。这些行为不仅损害了消费者权益,也破坏了正常的人伦关系和社会秩序。 第五,跨境金融服务监管协调有待加强。随着金融全球化的深入发展,跨境支付、数字资产交易等业务快速增长,但各国监管标准和执行尺度存在差异,容易产生监管套利和风险传染。一些失独家庭在境外寻求辅助生殖服务时,往往涉及复杂的跨境金融交易,但相关监管制度尚不完善。部分不法分子利用监管空白从事洗钱、恐怖主义融资等违法活动,甚至出现代孕弃养等道德风险,严重冲击了传统的生育权观念和家庭伦理。 从国际经验来看,欧盟《数字服务法》和《人工智能法案》为金融科技监管提供了重要参考。欧盟强调算法透明度和可解释性,要求高风险AI系统必须接受严格的合规性评估。美国金融稳定监督委员会发布的《金融科技风险监测框架》强调了系统性风险防范的重要性,建立了跨部门协调机制。新加坡金融管理局推出的"监管沙盒"机制,为金融创新提供了安全的试验环境,在控制风险的前提下允许金融机构测试创新产品和服务。 在具体制度设计上,建议采用"双轨制"监管模式:对于成熟稳定的金融科技业务,实行常规监管;对于新兴创新业务,设立专门的创新监管机制。同时,应当建立健全的消费者保护体系,加强投资者教育,提高公众的金融素养和风险意识。特别要关注社会弱势群体的金融需求,确保他们能够公平享受金融科技创新带来的便利。 监管科技(RegTech)的发展为提升监管效能提供了新的可能。通过运用大数据、区块链、人工智能等技术手段,监管部门可以实现实时监控、智能分析和精准执法。这不仅能够降低监管成本,还能提高监管的前瞻性和有效性。同时,监管科技也有助于及时发现和处置金融风险,防范系统性金融风险的发生。 在国际合作方面,应当积极参与全球金融科技监管标准的制定,推动建立统一的跨境监管协调机制。加强与主要经济体的监管对话与合作,共同应对金融科技发展带来的挑战。同时,要防范境外不法分子利用技术手段从事违法金融活动,维护国家金融安全。 展望未来,数字货币的发展将进一步推动金融业态的深刻变革。央行数字货币(CBDC)的研发和应用,将重塑货币政策传导机制和金融基础设施。虚拟资产、NFT等新兴数字资产的兴起,也对传统的资产管理和投资理念提出了新的挑战。监管部门需要前瞻性地研究相关政策框架,确保数字货币的安全稳定运行。 总而言之,构建适应数字经济发展的金融监管框架是一个系统工程,需要统筹考虑创新激励、风险防范、消费者保护、国际协调等多重目标。只有建立科学合理的制度体系,坚持以人为本的监管理念,才能真正实现金融科技的健康可持续发展,更好地服务实体经济和人民群众的金融需求,促进经济社会的和谐发展。
考点1: "监管套利" 必须译为 "regulatory arbitrage",固定的金融专业术语,指利用不同监管制度间的差异获取利益 考点2: "穿透式监管" 必须译为 "look-through supervision; penetrating supervision",监管领域专业术语,按业务实质进行监管 考点3: "监管沙盒" 必须译为 "regulatory sandbox",金融科技监管的固定专业术语 考点4: "央行数字货币" 必须译为 "Central Bank Digital Currency",国际通用的标准缩写和术语 考点5: "监管科技" 必须译为 "RegTech 或 Regulatory Technology)",行业标准缩写术语 考点6: "双轨制监管" 必须译为 "dual-track regulatory system",监管制度设计的专业术语 考点7: "功能监管" 必须译为 "functional regulation",与机构监管相对的监管模式专业术语 考点8: "行为监管" 必须译为 "conduct regulation",金融监管领域的固定术语 考点9: "宏观审慎监管" 必须译为 "macro-prudential regulation",金融稳定领域的标准术语 考点10: "微观审慎监管" 必须译为 "micro-prudential regulation",与宏观审慎相对的监管概念 考点11: "系统重要性金融机构"必须译为 "Systemically Important Financial Institutions",国际金融监管的专业术语 考点12: "反洗钱合规" 必须译为 "AML compliance 或 Anti-Money Laundering compliance)",合规领域的固定专业术语 考点13: "第三方支付" 必须译为 "third-party payment services",支付行业的固定术语 考点14: "影子银行" 必须译为 "shadow banking",金融体系中的专业概念
1600b
学术论文
社会科学
188
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 标题:从“默照”到“看话”:论禅宗机锋语言的“不立文字”与“以言遣言”之辩证 摘要: 禅宗,作为佛教在中国最独特的显现,其核心法门宣称“不立文字,教外别传”。然而吊诡的是,禅宗公案与语录中又充满了机锋、棒喝等极具表现力的语言运用。本文旨在剖析这一看似矛盾的现象,探究禅宗如何通过一种特殊的语言策略——“以言遣言”——来实现其“不立文字”的终极关怀。文章首先将回溯“不立文字”的本意,阐明其并非废弃言语,而是警惕语言作为“指月之指”的局限性。其次,本文将重点分析中唐以后“机锋”语言的兴起,以及宋代“看话禅”的成熟,是如何将语言从描述工具转变为一种能主动击碎思维定势的实践方法。通过对“公案”的参究和对“话头”的苦心孤诣,修行者被逼入“大疑情”的绝境,最终在言语道断处“打破漆桶”,豁然开悟。本文认为,禅宗的语言革命,恰恰体式现了其“见性成佛”之宗旨,即真正的超越,并非发生在语言之外,而是在语言的尽头。 正文: 一、“不立文字”的本怀:作为“指月之指”的语言 禅宗提出“不立文字,直指人心”,这常被误解为对一切文字和经典的彻底否定。实际上,禅宗并非主张缄默主义,而是深刻洞悉到语言的工具性与局限性。任何言教,无论多么精妙,都只是指向月亮的手指(“指月之指”),而非月亮本身。众生易犯“认指为月”的错误,即执着于经文的字面含义,而忘记其所指向的、超越言诠的实相本身。因此,“不立文字”的真正意涵在于,反对将佛法僵化为教条,强调个体必须亲身证悟,而非仅仅在名相概念中打转。语言是渡河的舟筏,一旦登岸,舟筏即应舍弃。这一思想构成了禅宗后续一切独特语言实践的哲学基础。 二、机锋与公案:从对话到实践的语言革命 中唐以降,禅宗大师们不再满足于平实的说法,转而发展出一种极其犀利、动态的对话方式——“机锋”。机锋往来,常常答非所问,充满悖论与突兀的动作(如棒喝),其目的并非传递信息,而是如电光石火般斩断学人的逻辑思路,使其惯常的思维模式瞬间短路。马祖道一的“即心即佛”与后来的“非心非佛”便是典型的例子,通过自我否定来破除学人对名相的执着。 到了宋代,这些充满张力的师徒对话被记录、整理下来,形成了“公案”。“公案”并非一个待解的逻辑谜题,而是一个供修行者“参究”的活生生的精神事件。它如同一面镜子,映照出修行者自身的思维葛藤。修行者必须全身心地投入公案之中,用整个生命去“碰壁”,而非用头脑去理解。 三、看话禅与大疑情:将语言推向极致的修行 在公案修行的基础上,大慧宗杲禅师正式确立了“看话禅”的法门。他教导修行者,不必理会公案的整个故事情节,只需抓住其中最关键、最无逻辑可循的一句话,即“话头”,并反复在心中追问。例如,面对“万法归一,一归何处?”这个公案,其“话头”便是“一归何处?”。又如,面对赵州禅师对“狗子有无佛性?”的回答“无”,其话头便是这个“无”字。 “看话头”的要旨在于,不能用逻辑去推理解释,而要将全部身心能量都凝聚在这个话头上,形成一个巨大而炽热的疑团——“大疑情”。这个疑情越是深重、坚固,就越能隔绝一切杂念。修行者日夜提撕,行住坐卧皆不离这个话头,直到有一天,因缘时节到来,疑团猛然爆破,如同“打破漆桶”,无明顿消,亲见本来面目。这一刻,语言的功能发挥到了极致,也在此刻彻底消亡,从而真正实现了“以言遣言”的超越。 结论:一种反身性的语言智慧 禅宗的“机锋”与“看话禅”,并非对“不立文字”的背离,反而是对其最深刻的贯彻。它展示了一种高超的反身性语言智慧:正因为认识到语言的局限,所以才要最大化地利用其“非逻辑”的潜能,将其锻造成一把可以刺破语言自身之网的利剑。这种“以言遣言”的辩证法,不仅是古代禅师的修行技艺,更为我们反思现代社会中语言与真实、符号与体验之间的关系,提供了一份无价的思想遗产。
标题:从“默照”到“看话”:论禅宗机锋语言的“不立文字”与“以言遣言”之辩证 摘要: 禅宗,作为佛教在中国最独特的显现,其核心法门宣称“不立文字,教外别传”。然而吊诡的是,禅宗公案与语录中又充满了机锋、棒喝等极具表现力的语言运用。本文旨在剖析这一看似矛盾的现象,探究禅宗如何通过一种特殊的语言策略——“以言遣言”——来实现其“不立文字”的终极关怀。文章首先将回溯“不立文字”的本意,阐明其并非废弃言语,而是警惕语言作为“指月之指”的局限性。其次,本文将重点分析中唐以后“机锋”语言的兴起,以及宋代“看话禅”的成熟,是如何将语言从描述工具转变为一种能主动击碎思维定势的实践方法。通过对“公案”的参究和对“话头”的苦心孤诣,修行者被逼入“大疑情”的绝境,最终在言语道断处“打破漆桶”,豁然开悟。本文认为,禅宗的语言革命,恰恰体式现了其“见性成佛”之宗旨,即真正的超越,并非发生在语言之外,而是在语言的尽头。 正文: 一、“不立文字”的本怀:作为“指月之指”的语言 禅宗提出“不立文字,直指人心”,这常被误解为对一切文字和经典的彻底否定。实际上,禅宗并非主张缄默主义,而是深刻洞悉到语言的工具性与局限性。任何言教,无论多么精妙,都只是指向月亮的手指(“指月之指”),而非月亮本身。众生易犯“认指为月”的错误,即执着于经文的字面含义,而忘记其所指向的、超越言诠的实相本身。因此,“不立文字”的真正意涵在于,反对将佛法僵化为教条,强调个体必须亲身证悟,而非仅仅在名相概念中打转。语言是渡河的舟筏,一旦登岸,舟筏即应舍弃。这一思想构成了禅宗后续一切独特语言实践的哲学基础。 二、机锋与公案:从对话到实践的语言革命 中唐以降,禅宗大师们不再满足于平实的说法,转而发展出一种极其犀利、动态的对话方式——“机锋”。机锋往来,常常答非所问,充满悖论与突兀的动作(如棒喝),其目的并非传递信息,而是如电光石火般斩断学人的逻辑思路,使其惯常的思维模式瞬间短路。马祖道一的“即心即佛”与后来的“非心非佛”便是典型的例子,通过自我否定来破除学人对名相的执着。 到了宋代,这些充满张力的师徒对话被记录、整理下来,形成了“公案”。“公案”并非一个待解的逻辑谜题,而是一个供修行者“参究”的活生生的精神事件。它如同一面镜子,映照出修行者自身的思维葛藤。修行者必须全身心地投入公案之中,用整个生命去“碰壁”,而非用头脑去理解。 三、看话禅与大疑情:将语言推向极致的修行 在公案修行的基础上,大慧宗杲禅师正式确立了“看话禅”的法门。他教导修行者,不必理会公案的整个故事情节,只需抓住其中最关键、最无逻辑可循的一句话,即“话头”,并反复在心中追问。例如,面对“万法归一,一归何处?”这个公案,其“话头”便是“一归何处?”。又如,面对赵州禅师对“狗子有无佛性?”的回答“无”,其话头便是这个“无”字。 “看话头”的要旨在于,不能用逻辑去推理解释,而要将全部身心能量都凝聚在这个话头上,形成一个巨大而炽热的疑团——“大疑情”。这个疑情越是深重、坚固,就越能隔绝一切杂念。修行者日夜提撕,行住坐卧皆不离这个话头,直到有一天,因缘时节到来,疑团猛然爆破,如同“打破漆桶”,无明顿消,亲见本来面目。这一刻,语言的功能发挥到了极致,也在此刻彻底消亡,从而真正实现了“以言遣言”的超越。 结论:一种反身性的语言智慧 禅宗的“机锋”与“看话禅”,并非对“不立文字”的背离,反而是对其最深刻的贯彻。它展示了一种高超的反身性语言智慧:正因为认识到语言的局限,所以才要最大化地利用其“非逻辑”的潜能,将其锻造成一把可以刺破语言自身之网的利剑。这种“以言遣言”的辩证法,不仅是古代禅师的修行技艺,更为我们反思现代社会中语言与真实、符号与体验之间的关系,提供了一份无价的思想遗产。
考点1:“不立文字”必须译为“no reliance on words and letters”或“no-dependence on words and letters” 考点2:“机锋”推荐译为“sharp verbal repartee” 考点3:“公案”必须译为“koan(s)”;“gongan (koans)” 考点4:“话头”必须译为“the critical phrase (Huatou)” 考点5:“见性成佛”必须译为“seeing one's true nature and attaining Buddhahood”或 “seeing one's nature and becoming a Buddha” 考点6:“以言遣言”推荐译为“using words to eliminate words”或 “using words to dispel words” 考点7:“大疑情”必须译为“the Great Doubt”或“great doubt” 考点8:“指月之指”推荐译为“the finger pointing at the moon”或;“a finger pointing at the moon” 考点9:“打破漆桶”必须译为“smashing the lacquer bucket”或 “breaking the lacquer bucket” 考点10:“默照禅”推荐译为“Silent Illumination Chan” 考点11:“绝境”推荐译为“desperate state” 考点12:“看话禅”必须译为“Word Contemplation Chan (Kanhua Chan)”
162a4
学术论文
人文科学
129
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Zonal flows, by which we mean azimuthally symmetric band-like shear flows, are a ubiquitous phenomena in nature and the laboratory. The well-known examples of the Jovian belts and zones, and the terrestrial atmospheric jet stream are familiar to nearly everyone—the latter especially to travellers enduring long, bumpy airplane rides against strong head winds. Zonal flows are also present in the Venusian atmosphere (which rotates faster than the planet does!) and occur in the solar tachocline, where they play a role in the solar dynamo mechanism. In the laboratory, the importance of sheared E × B flows to the development of L-mode confinement, the L-to-H transition and internal transport barriers(ITBs) is now well and widely appreciated. While many mechanisms can act to trigger and stimulate the growth of sheared electric fields (i.e. profile evolution and transport bifurcation, neoclassical effects, external momentum injection, etc) certainly one possibility is via the self-generation and amplification of E × B flows by turbulent stresses (i.e. the turbulent transport of momentum). Of course, this is the same mechanism as that responsible for zonal flow generation. It should be emphasized that it is now widely recognized and accepted that zonal flows are a key constituent in nearly all cases and regimes of drift wave turbulence—indeed, so much so that this classic problem is now frequently referred to as ‘drift wave–zonal flow turbulence’. This paradigm shift occurred on account of the realization that zonal flows are ubiquitous in dynamical models used for describing fusion plasmas (i.e. ITG, TEM, ETG, resistive ballooning and interchange, etc) in all geometries and regimes (i.e. core, edge, etc), and because of the realization that zonal flows are a critical agent of self-regulation for drift wave transport and turbulence. Both theoretical work and numerical simulation made important contributions to this paradigm shift. Indeed, for the case of low collisionality plasmas, a significant portion of the available free energy is ultimately deposited in the zonal flows. Figure 1 presents energy flow charts which illustrate the classic paradigm of drift wave turbulence and the new paradigm of drift wave–zonal flow turbulence. The study of zonal flow has had a profound impact on fusion research. For instance, the proper treatment of the zonal flow physics has resolved some of the confusion concerning the prospect of burning plasma, as has been discussed by Rosenbluth and collaborators in conjunction with the design of the International Thermonuclear Experimental Reactor (ITER). At the same time, the understanding of the turbulence–zonal flow system has advanced the understanding of self-organization processes in nature. We note here that, while zonal flows have a strong influence on the formation of transport barriers, the dynamics of barriers and transitions involve evolutions of both the mean E × B flow as well as the zonal E × B flow. The topics of mean Er dynamics, transport barriers and confinement regime transitions are beyond the scope of this review. In the context of tokamak plasmas, zonal flows are n = 0 electrostatic potential fluctuations with finite radial wavenumber. Zonal flows are elongated, asymmetric vortex modes, and thus have zero frequency. They are predominantly poloidally symmetric as well, though some coupling to low-m sideband modes may occur. On account of their symmetry, zonal flows cannot access expansion free energy stored in temperature, density gradients, etc, and are not subject to Landau damping. These zonal flows are driven exclusively by nonlinear interactions, which transfer energy from the finite-n drift waves to the n = 0 flow. Usually, such nonlinear interactions are three-wave triad couplings between two high k drift waves and one low q zonal flow excitation. In position space, this energy transfer process is simple one whereby Reynolds work is performed on the flow by the wave stresses. Two important consequences of this process of generation follow directly. First, since zonal flow production is exclusively via nonlinear transfer from drift waves, zonal flows must eventually decay and vanish if the underlying drift wave drive is extinguished. Thus, zonal flows differ in an important way from mean E × B flows, which can be sustained in the absence of turbulence (and are, in strong H-mode and ITB regimes). Second, since zonal flows are generated by nonlinear energy transfer from drift waves, their generation naturally acts to reduce the intensity and level of transport caused by the primary drift wave turbulence. Thus, zonal flows necessarily act to regulate and partially suppress drift wave turbulence and transport. This is clear from numerical simulations, which universally show that turbulence and transport levels are reduced when the zonal flow generation is (properly) allowed. Since zonal flows cannot tap expansion free energy, are generated by nonlinear coupling from drift waves, and damp primarily (but not exclusively) by collisional processes, they constitute a significant and benign (from a confinement viewpoint) reservoir or repository for the available free energy of the system. Another route to understanding the effects of zonal flow on drift waves is via the shearing paradigm. From this standpoint, zonal flows produce a spatio-temporally complex shearing pattern, which naturally tends to distort drift wave eddies by stretching them, and in the process generates large kr. Of course, at smaller scales, coupling to dissipation becomes stronger, resulting in a net stabilizing trend. The treatment of zonal flow shearing differs from that for mean flow shearing on account of the complexity of the flow pattern. Progress here has been facilitated by the realization that a statistical analysis is possible. This follows from the fact that the autocorrelation time of a drift wave-packet propagating in a zonal flow field is usually quite short, and because the drift wave rays are chaotic. Hence, significant advances have been made on calculating the ‘back reaction’ of zonal flows on the underlying drift wave field within the framework of random shearing, using wave kinetics and quasilinear theory. Conservation of energy between drift waves and zonal flows has been proved for the theory, at the level of a renormalized quasilinear description. Thus, it is possible to close the ‘feedback loop’ of wave–flow interactions, allowing a self-consistent analysis of the various system states, and enabling an understanding of the mechanisms and routes for bifurcation between them. Of course, it is patently obvious that the zonal flow problem is one of self-organization of a large structure in turbulence. Examples of other members of this class include transport barrier and profile formation and dynamics, the origin of the solar differential rotation, the famous magnetic dynamo problem (relevant, in quite different limits, to the sun, earth, galaxy and reversed field pinch), and the formation of profiles in turbulent and swirling pipe flow. Table 1 summarizes these related structure formation phenomena, illustrating the objective of this review. Most of these problems are attacked at the simplest level by considering the stability of an ensemble or ‘gas’ of ambient turbulence to a seed perturbation. For example, in the dynamo problem, one starts by considering the stability of some state of magnetohydrodynamic (MHD) turbulence to a seed magnetic field. In the zonal flow problem, one correspondingly considers the stability of a gas of drift waves to a seed shear. The incidence of instability means that the initial vortex tilt will be re-enforced, thus amplifying the seed perturbation. It should be noted that the zonal flow formation phenomenon is related to, but not quite the same as, the well-known inverse cascade of energy in a two-dimensional fluid, which leads to large scale vortex formation. This is because the inverse cascade proceeds via a local coupling in wavenumber space, while zonal flow generation occurs via non-local transfer of energy between small and large scales. Indeed, zonal shear amplification is rather like the familiar α-effect from dynamo theory, which describes a non-local transfer of magnetic helicity to large scale. We also note that the initial stage of pattern formation instability meets only part of the challenge to a theoretical description of structure formation, and that one must subsequently ‘close the loop’ by understanding the mechanisms of saturation of the zonal flow instability. The saturation of zonal flows driven by drift wave turbulence is now a subject of intensive theoretical and computational investigation, worldwide. As a related phenomena, convective cells have been subject to intensive study for a long time. The convective cell is a perturbation which is constant along the magnetic field line but changes in the direction perpendicular to the magnetic field. Such a structure is known to be induced by background drift wave turbulence. The zonal flow can be considered as a particular example of an anisotropic convective cell. However, the convective cells of greatest interest as agents of transport are localized in the poloidal direction and extended radially, which is the opposite limit of anisotropy from that of the zonal flow. Such cells are commonly referred to as streamers. As the zonal flow problem is a member of a large class of rapidly expanding research topics, the perspective of this review is composed as follows. First, we present detailed explanations of the physical understanding of drift wave–zonal flow turbulence. Second, we also stress the view that studies on toroidal plasma turbulence enhance our understanding of turbulent structure formation in nature. In this sense, this review is a companion paper to recent reviews on the magnetic dynamo problem which, taken together, present a unified view that addresses the mystery of structure formation in turbulent media. Third, the impact of direct nonlinear simulation (DNS) is discussed in the context of understanding zonal flow physics, although a survey of DNS techniques themselves is beyond the scope of this review. It is certainly the case that DNS studies have significantly furthered our understanding of drift wave–zonal flow turbulence. For these reasons, examples are mainly chosen from the realm of core plasma (i.e. drift wave) turbulence. In order to maintain transparency and to be concise, this review is limited in scope. Studies of edge turbulence and of general convective cell physics are not treated in depth here. While these topics are closely related to the topic of this review, extensive introductory discussions, which are too lengthy for this paper, are necessary. Hence, details of these important areas are left for future reviews.
Zonal flows, by which we mean azimuthally symmetric band-like shear flows, are a ubiquitous phenomena in nature and the laboratory. The well-known examples of the Jovian belts and zones, and the terrestrial atmospheric jet stream are familiar to nearly everyone—the latter especially to travellers enduring long, bumpy airplane rides against strong head winds. Zonal flows are also present in the Venusian atmosphere (which rotates faster than the planet does!) and occur in the solar tachocline, where they play a role in the solar dynamo mechanism. In the laboratory, the importance of sheared E × B flows to the development of L-mode confinement, the L-to-H transition and internal transport barriers(ITBs) is now well and widely appreciated. While many mechanisms can act to trigger and stimulate the growth of sheared electric fields (i.e. profile evolution and transport bifurcation, neoclassical effects, external momentum injection, etc) certainly one possibility is via the self-generation and amplification of E × B flows by turbulent stresses (i.e. the turbulent transport of momentum). Of course, this is the same mechanism as that responsible for zonal flow generation. It should be emphasized that it is now widely recognized and accepted that zonal flows are a key constituent in nearly all cases and regimes of drift wave turbulence—indeed, so much so that this classic problem is now frequently referred to as ‘drift wave–zonal flow turbulence’. This paradigm shift occurred on account of the realization that zonal flows are ubiquitous in dynamical models used for describing fusion plasmas (i.e. ITG, TEM, ETG, resistive ballooning and interchange, etc) in all geometries and regimes (i.e. core, edge, etc), and because of the realization that zonal flows are a critical agent of self-regulation for drift wave transport and turbulence. Both theoretical work and numerical simulation made important contributions to this paradigm shift. Indeed, for the case of low collisionality plasmas, a significant portion of the available free energy is ultimately deposited in the zonal flows. Figure 1 presents energy flow charts which illustrate the classic paradigm of drift wave turbulence and the new paradigm of drift wave–zonal flow turbulence. The study of zonal flow has had a profound impact on fusion research. For instance, the proper treatment of the zonal flow physics has resolved some of the confusion concerning the prospect of burning plasma, as has been discussed by Rosenbluth and collaborators in conjunction with the design of the International Thermonuclear Experimental Reactor (ITER). At the same time, the understanding of the turbulence–zonal flow system has advanced the understanding of self-organization processes in nature. We note here that, while zonal flows have a strong influence on the formation of transport barriers, the dynamics of barriers and transitions involve evolutions of both the mean E × B flow as well as the zonal E × B flow. The topics of mean Er dynamics, transport barriers and confinement regime transitions are beyond the scope of this review. In the context of tokamak plasmas, zonal flows are n = 0 electrostatic potential fluctuations with finite radial wavenumber. Zonal flows are elongated, asymmetric vortex modes, and thus have zero frequency. They are predominantly poloidally symmetric as well, though some coupling to low-m sideband modes may occur. On account of their symmetry, zonal flows cannot access expansion free energy stored in temperature, density gradients, etc, and are not subject to Landau damping. These zonal flows are driven exclusively by nonlinear interactions, which transfer energy from the finite-n drift waves to the n = 0 flow. Usually, such nonlinear interactions are three-wave triad couplings between two high k drift waves and one low q zonal flow excitation. In position space, this energy transfer process is simple one whereby Reynolds work is performed on the flow by the wave stresses. Two important consequences of this process of generation follow directly. First, since zonal flow production is exclusively via nonlinear transfer from drift waves, zonal flows must eventually decay and vanish if the underlying drift wave drive is extinguished. Thus, zonal flows differ in an important way from mean E × B flows, which can be sustained in the absence of turbulence (and are, in strong H-mode and ITB regimes). Second, since zonal flows are generated by nonlinear energy transfer from drift waves, their generation naturally acts to reduce the intensity and level of transport caused by the primary drift wave turbulence. Thus, zonal flows necessarily act to regulate and partially suppress drift wave turbulence and transport. This is clear from numerical simulations, which universally show that turbulence and transport levels are reduced when the zonal flow generation is (properly) allowed. Since zonal flows cannot tap expansion free energy, are generated by nonlinear coupling from drift waves, and damp primarily (but not exclusively) by collisional processes, they constitute a significant and benign (from a confinement viewpoint) reservoir or repository for the available free energy of the system. Another route to understanding the effects of zonal flow on drift waves is via the shearing paradigm. From this standpoint, zonal flows produce a spatio-temporally complex shearing pattern, which naturally tends to distort drift wave eddies by stretching them, and in the process generates large kr. Of course, at smaller scales, coupling to dissipation becomes stronger, resulting in a net stabilizing trend. The treatment of zonal flow shearing differs from that for mean flow shearing on account of the complexity of the flow pattern. Progress here has been facilitated by the realization that a statistical analysis is possible. This follows from the fact that the autocorrelation time of a drift wave-packet propagating in a zonal flow field is usually quite short, and because the drift wave rays are chaotic. Hence, significant advances have been made on calculating the ‘back reaction’ of zonal flows on the underlying drift wave field within the framework of random shearing, using wave kinetics and quasilinear theory. Conservation of energy between drift waves and zonal flows has been proved for the theory, at the level of a renormalized quasilinear description. Thus, it is possible to close the ‘feedback loop’ of wave–flow interactions, allowing a self-consistent analysis of the various system states, and enabling an understanding of the mechanisms and routes for bifurcation between them. Of course, it is patently obvious that the zonal flow problem is one of self-organization of a large structure in turbulence. Examples of other members of this class include transport barrier and profile formation and dynamics, the origin of the solar differential rotation, the famous magnetic dynamo problem (relevant, in quite different limits, to the sun, earth, galaxy and reversed field pinch), and the formation of profiles in turbulent and swirling pipe flow. Table 1 summarizes these related structure formation phenomena, illustrating the objective of this review. Most of these problems are attacked at the simplest level by considering the stability of an ensemble or ‘gas’ of ambient turbulence to a seed perturbation. For example, in the dynamo problem, one starts by considering the stability of some state of magnetohydrodynamic (MHD) turbulence to a seed magnetic field. In the zonal flow problem, one correspondingly considers the stability of a gas of drift waves to a seed shear. The incidence of instability means that the initial vortex tilt will be re-enforced, thus amplifying the seed perturbation. It should be noted that the zonal flow formation phenomenon is related to, but not quite the same as, the well-known inverse cascade of energy in a two-dimensional fluid, which leads to large scale vortex formation. This is because the inverse cascade proceeds via a local coupling in wavenumber space, while zonal flow generation occurs via non-local transfer of energy between small and large scales. Indeed, zonal shear amplification is rather like the familiar α-effect from dynamo theory, which describes a non-local transfer of magnetic helicity to large scale. We also note that the initial stage of pattern formation instability meets only part of the challenge to a theoretical description of structure formation, and that one must subsequently ‘close the loop’ by understanding the mechanisms of saturation of the zonal flow instability. The saturation of zonal flows driven by drift wave turbulence is now a subject of intensive theoretical and computational investigation, worldwide. As a related phenomena, convective cells have been subject to intensive study for a long time. The convective cell is a perturbation which is constant along the magnetic field line but changes in the direction perpendicular to the magnetic field. Such a structure is known to be induced by background drift wave turbulence. The zonal flow can be considered as a particular example of an anisotropic convective cell. However, the convective cells of greatest interest as agents of transport are localized in the poloidal direction and extended radially, which is the opposite limit of anisotropy from that of the zonal flow. Such cells are commonly referred to as streamers. As the zonal flow problem is a member of a large class of rapidly expanding research topics, the perspective of this review is composed as follows. First, we present detailed explanations of the physical understanding of drift wave–zonal flow turbulence. Second, we also stress the view that studies on toroidal plasma turbulence enhance our understanding of turbulent structure formation in nature. In this sense, this review is a companion paper to recent reviews on the magnetic dynamo problem which, taken together, present a unified view that addresses the mystery of structure formation in turbulent media. Third, the impact of direct nonlinear simulation (DNS) is discussed in the context of understanding zonal flow physics, although a survey of DNS techniques themselves is beyond the scope of this review. It is certainly the case that DNS studies have significantly furthered our understanding of drift wave–zonal flow turbulence. For these reasons, examples are mainly chosen from the realm of core plasma (i.e. drift wave) turbulence. In order to maintain transparency and to be concise, this review is limited in scope. Studies of edge turbulence and of general convective cell physics are not treated in depth here. While these topics are closely related to the topic of this review, extensive introductory discussions, which are too lengthy for this paper, are necessary. Hence, details of these important areas are left for future reviews.
考点1:”zonal flow“必须译为”带状流“; 考点2:”drift waveturbulence transport“必须译为”漂移波湍流输运“; 考点3:”Reynolds stress“必须译为”雷诺应力“; 考点4:”inverse cascade of energy“必须译为”能量逆级联“; 考点5:”magnetohydrodynamic“必须译为”磁流体动力学“; 考点6:”azimuthally symmetric“必须译为”轴对称“; 考点7:”convective cells“必须译为”对流元“; 考点8:”Jovian belts“必须译为”木星云带“; 考点9:”regimes of drift wave turbulence“必须译为”漂移波湍流(参数)区域“; 考点10:”Rosenbluth“人名不推荐翻译为中文; 考点11:”underlying“推荐译为”潜在的“; 考点12:”streamers“必须译为”流带“; 考点13:”differ in an important way from“推荐译为”重要区别是“。
16ac7
学术论文
自然科学
145
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 北京孔庙始建于元大德六年(1302),大德十年(1306)建成,明永乐九年(1411)重建。北京孔庙从1928年起对普通人外开放,以前都是皇帝及其随员供人游览才可以进入。北京孔庙的神圣性来自于其严格按照礼制要求建造,而这种礼制要求体现在在不同的尺度中,也就是说,孔庙的神圣性是在不同尺度中被建构的。 孔庙的选址十分强调讲究中国的风水。在选址时遵循风水的基本规则意味着一方面,可以在客观上保证孔庙的采光、通风和保温等能达到最好的效果;风水有两个流派,一个是依据自然环境选址,主要流行于山区;另一是依据方位选址,主要流行在平原方面。北京城坐落在平原,因此适合采用按照方位选址的风水理论。文庙的选址,其目的是,也可以求孔子和诸位儒学先贤保佑人们满足人们对科举通过科举走上仕途。按照方位选址主要看地面与天上的星辰的对应关系。“庙者天星也……上应天上之星辰,下司人间之福禄。居凶方则凶,居吉方则吉。吉者,三吉六秀是也。亥震庚,为三吉。艮丙巽辛兑丁,为六秀。庙居此方,官民俱吉。” 历史上甚至有因为风水不佳而将孔庙迁址的例子。《阳宅三要》中说“阴阳之理、自古攸分,二者不和,凶气必至。故公衙要合法,而庙亦不可不居乎吉地......文庙建于甲、艮、巽三地为得地也......”《相宅经纂》中也有“文庙建甲、艮、巽三方,为得地。庙后易高耸,如笔如抢,左宜空缺明亮,一眼看见奎文楼,大利科甲。”的记录。(甲为震,东方;艮为东北;巽为东南。)由此可见东、东北和东南是文庙的常见选址。元朝初建北京孔庙,选址就位于皇城的东北方(图5-1)。从元大都到明清皇城再到今天的老城,北京核心区固然经历了改造,但是其基本格局还是得到了保留(图5)。因此,如今的孔庙依然位于城市中心的东北方。这种选址体现了孔庙的地位,塑造了其神圣性。 孔庙所在的街道设计布局为孔庙的神圣感提供了三层烘托铺垫。第一,孔庙位于成贤街(今国子监街)。东西两侧的街口分别立有一彩绘牌坊,上书“成贤街”。这两个牌坊因其色彩艳丽、形制高大而格外醒目,因此也提醒进入街道的人们集中注意力开始关注街道内的景观。第二,由东口进入街道,则很快可以看见刻有“文武官员至此下马”的下马碑(位于孔庙东侧);由西口进入,则可以见到街道内的第二个彩绘牌坊——“国子监”牌坊(位于国子监西侧)。这两个建筑都能够再次提醒游客自己已经走近神圣空间,需要更加注意自己的行为礼仪。第三,走到孔庙与国子监大门(先师门)对面博物馆门口,可以看见是一座琉璃瓦的一字影壁(建于1501年)。和宏伟的影壁和先师门。这两个建筑都十分高大,意在比喻孔子的学识道德之高深,勉励学子勤学。它们的存在使得人们游客一来意识到自己已经走到孔庙跟前,而来使得人们无法一眼看见孔庙内部,由此对内部建筑的神秘产生好奇和崇敬。 清代,自康熙皇帝开始,朝廷重视孔庙周边的管理。清代是满族人建立的王朝,满族人从中国东北进入中原之前不信儒学,因此到清代第三位皇帝将都城迁到北京后,并不将汉族人信奉的孔庙当做神圣空间。直到清代第四位皇帝康熙执政期间,才开始重视孔庙。这是因为这时满族军队基本上控制了中国绝大部分地区,康熙皇帝开始考虑处理满族和汉族的关系。顺天府是首都北京下辖的两个府之一。1682年的一天,康熙皇帝看到顺天府的官员提交的一份报告。报告说学宫门摆放马槽,栓着军队一些官员的马匹,道路上到处是马粪马尿。这些官员是满族军队中级别最高的镶黄旗总指挥的部下。这些满足官人在学宫院内挖土做砖,使院子里有一大坑,院子四周的墙也被破坏。孔庙内还养着牛羊。这些做法都是对神圣空间的不尊敬。康熙皇帝看到这个报告后,马上命令负责国家礼仪的部门将院墙修好,不允许在此养马、牛、羊。最后,康熙皇帝还表扬了这位提交报告汉族官员(K. Li, 1912),由此缓和满族和汉族的文化对立。1684年,康熙为大成殿题写“万世师表”(永远是人们的榜样)的牌匾。
北京孔庙始建于元大德六年(1302),大德十年(1306)建成,明永乐九年(1411)重建。北京孔庙从1928年起对普通人外开放,以前都是皇帝及其随员供人游览才可以进入。北京孔庙的神圣性来自于其严格按照礼制要求建造,而这种礼制要求体现在在不同的尺度中,也就是说,孔庙的神圣性是在不同尺度中被建构的。 孔庙的选址十分强调讲究中国的风水。在选址时遵循风水的基本规则意味着一方面,可以在客观上保证孔庙的采光、通风和保温等能达到最好的效果;风水有两个流派,一个是依据自然环境选址,主要流行于山区;另一是依据方位选址,主要流行在平原方面。北京城坐落在平原,因此适合采用按照方位选址的风水理论。文庙的选址,其目的是,也可以求孔子和诸位儒学先贤保佑人们满足人们对科举通过科举走上仕途。按照方位选址主要看地面与天上的星辰的对应关系。“庙者天星也……上应天上之星辰,下司人间之福禄。居凶方则凶,居吉方则吉。吉者,三吉六秀是也。亥震庚,为三吉。艮丙巽辛兑丁,为六秀。庙居此方,官民俱吉。” 历史上甚至有因为风水不佳而将孔庙迁址的例子。《阳宅三要》中说“阴阳之理、自古攸分,二者不和,凶气必至。故公衙要合法,而庙亦不可不居乎吉地......文庙建于甲、艮、巽三地为得地也......”《相宅经纂》中也有“文庙建甲、艮、巽三方,为得地。庙后易高耸,如笔如抢,左宜空缺明亮,一眼看见奎文楼,大利科甲。”的记录。(甲为震,东方;艮为东北;巽为东南。)由此可见东、东北和东南是文庙的常见选址。元朝初建北京孔庙,选址就位于皇城的东北方(图5-1)。从元大都到明清皇城再到今天的老城,北京核心区固然经历了改造,但是其基本格局还是得到了保留(图5)。因此,如今的孔庙依然位于城市中心的东北方。这种选址体现了孔庙的地位,塑造了其神圣性。 孔庙所在的街道设计布局为孔庙的神圣感提供了三层烘托铺垫。第一,孔庙位于成贤街(今国子监街)。东西两侧的街口分别立有一彩绘牌坊,上书“成贤街”。这两个牌坊因其色彩艳丽、形制高大而格外醒目,因此也提醒进入街道的人们集中注意力开始关注街道内的景观。第二,由东口进入街道,则很快可以看见刻有“文武官员至此下马”的下马碑(位于孔庙东侧);由西口进入,则可以见到街道内的第二个彩绘牌坊——“国子监”牌坊(位于国子监西侧)。这两个建筑都能够再次提醒游客自己已经走近神圣空间,需要更加注意自己的行为礼仪。第三,走到孔庙与国子监大门(先师门)对面博物馆门口,可以看见是一座琉璃瓦的一字影壁(建于1501年)。和宏伟的影壁和先师门。这两个建筑都十分高大,意在比喻孔子的学识道德之高深,勉励学子勤学。它们的存在使得人们游客一来意识到自己已经走到孔庙跟前,而来使得人们无法一眼看见孔庙内部,由此对内部建筑的神秘产生好奇和崇敬。 清代,自康熙皇帝开始,朝廷重视孔庙周边的管理。清代是满族人建立的王朝,满族人从中国东北进入中原之前不信儒学,因此到清代第三位皇帝将都城迁到北京后,并不将汉族人信奉的孔庙当做神圣空间。直到清代第四位皇帝康熙执政期间,才开始重视孔庙。这是因为这时满族军队基本上控制了中国绝大部分地区,康熙皇帝开始考虑处理满族和汉族的关系。顺天府是首都北京下辖的两个府之一。1682年的一天,康熙皇帝看到顺天府的官员提交的一份报告。报告说学宫门摆放马槽,栓着军队一些官员的马匹,道路上到处是马粪马尿。这些官员是满族军队中级别最高的镶黄旗总指挥的部下。这些满足官人在学宫院内挖土做砖,使院子里有一大坑,院子四周的墙也被破坏。孔庙内还养着牛羊。这些做法都是对神圣空间的不尊敬。康熙皇帝看到这个报告后,马上命令负责国家礼仪的部门将院墙修好,不允许在此养马、牛、羊。最后,康熙皇帝还表扬了这位提交报告汉族官员(K. Li, 1912),由此缓和满族和汉族的文化对立。1684年,康熙为大成殿题写“万世师表”(永远是人们的榜样)的牌匾。
考点1:“文庙”必须翻译为“the Temple of Confucius / Confucian Temple” 考点2:“庙象天星”推荐翻译为“the temple corresponds to celestial stars”(含义对即可) 考点3:“公衙要合法”推荐翻译为“the location of the public office should be reasonable”(含义对即可) 考点4:“三吉”推荐翻译为“San Ji” 考点5:“六秀”推荐翻译为“Liu Xiu” 考点6:“阳宅三要”推荐翻译为“The Three Essentials of Yang Residence” 考点7:“相宅经纂”推荐翻译为“Classic for Assessing House” 考点8:“科甲”推荐翻译为“imperial examination success”(含义对即可) 考点9:“庙后易高耸,如笔如抢”推荐翻译为“The back of the temple should have a towering building, as straight as a pen, as prominent as a snatch.”(含义对即可) 考点10:“左宜空缺明亮”推荐翻译为“bright and open left side”(含义对即可) 考点11:“彩绘牌坊”推荐翻译为“painted memorial arch/paifang/Pailou” 考点12:“文武官员至此下马”推荐翻译为“civil and military officials dismounted here”(含义对即可) 考点13:“琉璃瓦的一字影壁”推荐翻译为“glazed shadow wall” 考点14:“官员”推荐翻译为“governor” 考点15:“镶黄旗”必须翻译为“Xianghuang settlement” 考点16:“万世师表”推荐翻译为“the eternal paragon” 考点17:“元大都”推荐翻译为“Yuan capital”
176ea
学术论文
人文科学
128
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Concrete is a composite construction material composed of a cementitious binder (usually Portland cement), water, and aggregates (sand and gravel or crushed stone), along with various admixtures that may be added to modify its properties. When these ingredients are first mixed together, the fresh concrete is a viscoplastic, moldable mixture that can be placed into forms. A series of chemical reactions called cement hydration begins as soon as water contacts the cement particles. Through hydration, the cement paste gradually hardens and gains strength, binding the aggregates together into a solid mass. This transformation from a fluid plastic state to a rigid, load-bearing solid typically occurs within a few hours after mixing (initial setting often happens after around 2 hours for a normal mix). Strength development then continues over time as curing progresses – concrete keeps gaining strength for months or even years, although the rate of strength gain diminishes with time. In practice, the compressive strength measured at 28 days (the 28-day strength) is commonly used as a benchmark of concrete quality (denoted as f’c in many specifications). About 90% of the long-term strength is usually achieved in the first month under proper curing conditions. Proper curing during this period is critical to concrete quality: the concrete must be prevented from drying out or freezing at early ages. Newly placed concrete is typically kept moist (for example by frequent water spraying or by covering with wet burlap and plastic sheets) and maintained at a suitable temperature. Such curing ensures that hydration can progress and that the concrete’s surface does not crack or lose strength prematurely. If concrete is allowed to dry too soon, it will stop gaining strength and may develop surface shrinkage cracks; if it freezes before sufficient hardening, its internal structure can be disrupted. Therefore, curing measures are often applied for at least 7–14 days to promote good strength and durability development. The strength of hardened concrete depends on many factors, including the mix proportions, the properties of the constituent materials, the curing regime, and the presence of any admixtures or additives. One of the most important factors is the water-to-cement ratio (w/c ratio). This is the ratio of the mass of water to the mass of cement in the mix. A lower w/c ratio generally leads to higher strength and density (since less excess water leaves behind fewer pores in the hardened concrete), but if the w/c is too low, the mix may become unworkable (too stiff and difficult to place). For complete hydration of cement, a minimum w/c of around 0.22–0.25 by weight is needed, but typical structural concrete mixes use a w/c in the range of 0.4 to 0.6 to ensure adequate workability. Workability refers to the ease with which fresh concrete can be mixed, placed, compacted, and finished without segregation. To improve workability without adding excess water (which would reduce strength and increase permeability), chemical admixtures such as plasticizers (water-reducing agents) are often used. For example, a high-range water reducer (superplasticizer) can produce a highly flowable concrete (even self-leveling) at a w/c as low as ~0.30, enabling very high strengths while still allowing the mix to be placed easily. A highly workable mix can flow and fill formwork and surround dense reinforcement with relative ease, but if too much water or improper proportions are used, the mix may segregate or bleed. Segregation is the separation of coarse aggregate from the mortar (cement paste and fines) due to settling, resulting in a non-uniform composition. Bleeding is a related phenomenon where excess water rises to the top surface of the concrete before it sets. Both segregation and bleeding indicate an unstable mix; they can cause weak, porous regions and lead to finishing problems or reduced durability in the hardened concrete. Proper mix proportioning and the use of fines and viscosity-modifying admixtures can minimize these issues. The workability of fresh concrete is commonly measured by the slump test, which provides an empirical indication of concrete consistency. In a standard slump test, fresh concrete is placed into a frustum-shaped cone mold (12 inches high) in three layers, and each layer is rodded (tamped) a specified number of times. The mold is then carefully lifted off, and the unsupported concrete slumps under its own weight. The vertical drop of the concrete’s top (measured relative to the mold height) is called the slump. A higher slump (e.g. 7 or 8 inches) indicates a very wet or fluid mix, while a low slump (e.g. 1 inch) indicates a dry, stiff mix. Typical concretes for general construction have slumps in the range of about 2 to 4 inches (50–100 mm) for a good balance of workability and strength. However, slump does not capture all aspects of workability; different mixes can have the same slump yet handle very differently (for instance, one may segregate more easily than another). Other tests and measures (such as flow table tests for mortar, the compaction factor test, or more advanced rheometer tests) can characterize workability, especially for specialized concretes. Still, the slump test remains the simplest and most widely used field test to verify batch-to-batch consistency of concrete and ensure that the fresh mix has not changed. In recent years, self-consolidating concrete (SCC) has been developed as an extremely high-workability mix that can flow and compact under its own weight without vibration. SCC is achieved by using advanced admixtures (superplasticizers) and viscosity-modifying agents to produce a fluid yet stable mix. It demonstrates that with appropriate mix design, one can obtain both high workability and high strength (through a low w/c ratio), which is valuable for placing concrete in structures with congested reinforcement or complex formwork. As the fresh concrete gradually sets and hardens, its properties transition to those of the hardened material. The density of normal-weight concrete is around 2400 kg/m³ (about 150 lb/ft³), and its compressive strength can range widely. Ordinary structural concrete typically achieves a 28-day compressive strength on the order of 20–40 MPa (3000–6000 psi), whereas high-strength concrete used in high-rise buildings or other special applications can reach strengths of 70 MPa (10,000 psi) or more. Concrete’s tensile strength is much lower than its compressive strength (on the order of only 10% of the compressive strength). For this reason, concrete is usually reinforced with steel bars (reinforcing steel, or “rebar”) to carry tension. The stress–strain curve of concrete in compression is nonlinear – unlike steel, concrete does not have a long elastic range or a well-defined yield point. It initially deforms approximately elastically at very low stress, but as stress increases, the curve gradually bends (due to microcracking and increasing inelasticity) and eventually reaches a peak, followed by a drop-off as the concrete fails by crushing. Because there is no sharply defined elastic limit, the modulus of elasticity of concrete is usually defined either as an initial tangent modulus or a secant modulus to a point on the curve (for example, to 50% of ultimate strength). For typical normal-strength concrete, a common empirical approximation is that the Young’s modulus is about 20–30 GPa (gigapascals), which corresponds to around 3 to 5 million psi. Higher-strength concretes have higher stiffness, though not in direct proportion to their strength. Volume stability is another important aspect of concrete performance. As concrete cures and dries, it undergoes shrinkage – a reduction in volume. Drying shrinkage occurs as water that is not chemically bound in hydration products gradually evaporates from the hardened concrete. If a large element (like a long slab or wall) is restrained while it is shrinking (for example, by the subgrade or by adjacent structural elements), tensile stresses can develop, potentially leading to cracking. To control crack formation, engineers often incorporate contraction joints (planned grooves or weak planes where shrinkage cracks can form in a controlled way) and provide sufficient distributed reinforcement (sometimes called temperature and shrinkage steel) to hold any cracks tightly closed. Thermal changes can also cause concrete to expand or contract. For instance, the heat released during cement hydration can raise the temperature of a large concrete mass in early-age, and subsequent cooling can induce tensile stresses. In hot weather, rapid surface drying and cooling of freshly placed concrete can cause a surface contraction while the interior is still warm, sometimes leading to a pattern of shallow cracks called thermal cracking. Measures such as gradual cooling of massive pours, insulating blankets, or using low-heat cements help mitigate these thermal stresses. In addition to shrinkage and thermal effects, concrete exhibits creep, which is the long-term gradual deformation under sustained load. Under a constant compressive stress (such as a heavy dead load on a column or beam), concrete will slowly deform over months and years beyond the initial elastic strain – this time-dependent strain is creep. Creep strain can be significant, sometimes on the same order as the initial elastic strain. It is influenced by factors like the concrete’s compressive strength, mix composition, age at loading, and the magnitude of the sustained stress (relative to strength). In structural design, creep is taken into account because it can increase deflections of beams over time and can cause redistribution of internal forces in statically indeterminate structures. However, keeping sustained stresses low (e.g. by using a higher concrete strength or adding extra reinforcement) can limit the magnitude of creep deformation. Durability is a critical aspect of civil materials engineering, and concrete must be designed to withstand the environmental conditions to which it will be exposed over decades. A key durability concern in cold climates is resistance to freeze–thaw cycles. When porous, saturated concrete freezes, the water in its pores expands about 9% in volume, which can create internal stresses that crack the concrete. Repeated freezing and thawing can cause progressive damage such as surface scaling and loss of strength. To improve freeze–thaw durability, modern concrete often includes air-entraining admixtures that create a network of microscopic air bubbles in the hardened concrete. These tiny air voids (~0.1 mm in size, with a proper spacing factor) act as pressure relief chambers for freezing water, thereby reducing internal stress. Air-entrained concrete can survive hundreds of freeze–thaw cycles without significant deterioration, whereas non-entrained concrete (especially if it has a high w/c ratio and hence high saturation) may suffer severe damage after just a few winters. Another long-term durability issue is chemical attack. For instance, sulfate attack occurs when sulfate ions (from sulfate-rich soils, groundwater, or seawater) penetrate the concrete and react with components of the cement paste (like calcium aluminate hydrates), forming expansive products such as ettringite and gypsum. These expansions can disrupt the concrete’s matrix, causing cracking and loss of cohesion over time. To combat sulfate attack, engineers can use sulfate-resistant cement (with low tricalcium aluminate content), incorporate pozzolanic materials, and ensure the concrete has a low permeability (achieved by a low w/c ratio and adequate curing). Similarly, a deleterious reaction called alkali–silica reaction (ASR) can occur between alkali hydroxides in cement paste and certain reactive silica minerals in some aggregates. ASR produces an expansive gel that swells as it absorbs water, leading to internal pressures and a pattern of map cracking in the concrete over years. To prevent ASR, engineers may use low-alkali cement and incorporate supplementary cementitious materials (like fly ash or slag) to tie up alkalis and dilute the reactive constituents, and they avoid using known reactive aggregates when possible. Moreover, many modern concrete mixtures include supplementary cementitious materials (SCMs) such as fly ash (a by-product from coal power plants), ground granulated blast-furnace slag, and silica fume. These SCMs participate in secondary pozzolanic reactions with the cement hydration products, refining the pore structure of the concrete and often improving its long-term performance. Concretes with SCMs tend to have lower permeability and higher later-age strength, which enhances durability by reducing the ingress of harmful agents. For example, adding fly ash or slag can significantly mitigate sulfate attack and alkali–silica reaction, while silica fume is well known for increasing strength and making the concrete matrix denser. The inclusion of SCMs also has sustainability benefits by recycling industrial by-products and reducing the required amount of Portland cement. Furthermore, reinforcement corrosion is a major threat to reinforced concrete structures, especially in harsh environments. Steel rebar embedded in concrete is normally protected by the high alkalinity of cement paste (which forms a thin passive oxide layer on the steel surface). However, if chloride ions (for example, from deicing salts or marine spray) penetrate through concrete and reach the steel, they can break down this passive layer and cause corrosion of the rebar. The corrosion products (iron oxides and hydroxides) occupy a greater volume than the original steel, exerting expansive pressure inside the concrete. This leads to cracking and spalling of the concrete cover, further accelerating deterioration. To improve durability against reinforcement corrosion, concrete can be made with low permeability (to slow chloride ingress) by using a low w/c ratio and SCMs; sufficient concrete cover thickness over the steel is provided to lengthen the travel path for chlorides; and sometimes corrosion-inhibiting admixtures or epoxy-coated (or galvanized) rebars are used. Good construction practices, such as adequate curing and avoiding any premature drying or thermal shock, also help produce a dense, impermeable concrete less vulnerable to these forms of attack. In summary, civil engineering materials like concrete must be designed not only for initial strength and workability but also for long-term durability. A well-proportioned concrete mix with an appropriate water–cement ratio, good-quality aggregates, and suitable admixtures can achieve the required strength while remaining sufficiently workable for placement. Through proper curing and quality control, the concrete will develop the desired mechanical properties (high compressive strength, adequate modulus, etc.) and a dense microstructure that resists deleterious agents. Attention to potential deterioration mechanisms – such as freeze–thaw damage, sulfate or chemical attack, ASR, and steel corrosion – is essential to ensure that structures made of concrete remain safe and serviceable throughout their intended life span. Ongoing advancements in materials engineering continue to improve concrete technology, yielding new types such as high-performance concrete (HPC) and fiber-reinforced concrete, which offer enhanced workability, strength, and durability for the infrastructure of the future.
Concrete is a composite construction material composed of a cementitious binder (usually Portland cement), water, and aggregates (sand and gravel or crushed stone), along with various admixtures that may be added to modify its properties. When these ingredients are first mixed together, the fresh concrete is a viscoplastic, moldable mixture that can be placed into forms. A series of chemical reactions called cement hydration begins as soon as water contacts the cement particles. Through hydration, the cement paste gradually hardens and gains strength, binding the aggregates together into a solid mass. This transformation from a fluid plastic state to a rigid, load-bearing solid typically occurs within a few hours after mixing (initial setting often happens after around 2 hours for a normal mix). Strength development then continues over time as curing progresses – concrete keeps gaining strength for months or even years, although the rate of strength gain diminishes with time. In practice, the compressive strength measured at 28 days (the 28-day strength) is commonly used as a benchmark of concrete quality (denoted as f’c in many specifications). About 90% of the long-term strength is usually achieved in the first month under proper curing conditions. Proper curing during this period is critical to concrete quality: the concrete must be prevented from drying out or freezing at early ages. Newly placed concrete is typically kept moist (for example by frequent water spraying or by covering with wet burlap and plastic sheets) and maintained at a suitable temperature. Such curing ensures that hydration can progress and that the concrete’s surface does not crack or lose strength prematurely. If concrete is allowed to dry too soon, it will stop gaining strength and may develop surface shrinkage cracks; if it freezes before sufficient hardening, its internal structure can be disrupted. Therefore, curing measures are often applied for at least 7–14 days to promote good strength and durability development. The strength of hardened concrete depends on many factors, including the mix proportions, the properties of the constituent materials, the curing regime, and the presence of any admixtures or additives. One of the most important factors is the water-to-cement ratio (w/c ratio). This is the ratio of the mass of water to the mass of cement in the mix. A lower w/c ratio generally leads to higher strength and density (since less excess water leaves behind fewer pores in the hardened concrete), but if the w/c is too low, the mix may become unworkable (too stiff and difficult to place). For complete hydration of cement, a minimum w/c of around 0.22–0.25 by weight is needed, but typical structural concrete mixes use a w/c in the range of 0.4 to 0.6 to ensure adequate workability. Workability refers to the ease with which fresh concrete can be mixed, placed, compacted, and finished without segregation. To improve workability without adding excess water (which would reduce strength and increase permeability), chemical admixtures such as plasticizers (water-reducing agents) are often used. For example, a high-range water reducer (superplasticizer) can produce a highly flowable concrete (even self-leveling) at a w/c as low as ~0.30, enabling very high strengths while still allowing the mix to be placed easily. A highly workable mix can flow and fill formwork and surround dense reinforcement with relative ease, but if too much water or improper proportions are used, the mix may segregate or bleed. Segregation is the separation of coarse aggregate from the mortar (cement paste and fines) due to settling, resulting in a non-uniform composition. Bleeding is a related phenomenon where excess water rises to the top surface of the concrete before it sets. Both segregation and bleeding indicate an unstable mix; they can cause weak, porous regions and lead to finishing problems or reduced durability in the hardened concrete. Proper mix proportioning and the use of fines and viscosity-modifying admixtures can minimize these issues. The workability of fresh concrete is commonly measured by the slump test, which provides an empirical indication of concrete consistency. In a standard slump test, fresh concrete is placed into a frustum-shaped cone mold (12 inches high) in three layers, and each layer is rodded (tamped) a specified number of times. The mold is then carefully lifted off, and the unsupported concrete slumps under its own weight. The vertical drop of the concrete’s top (measured relative to the mold height) is called the slump. A higher slump (e.g. 7 or 8 inches) indicates a very wet or fluid mix, while a low slump (e.g. 1 inch) indicates a dry, stiff mix. Typical concretes for general construction have slumps in the range of about 2 to 4 inches (50–100 mm) for a good balance of workability and strength. However, slump does not capture all aspects of workability; different mixes can have the same slump yet handle very differently (for instance, one may segregate more easily than another). Other tests and measures (such as flow table tests for mortar, the compaction factor test, or more advanced rheometer tests) can characterize workability, especially for specialized concretes. Still, the slump test remains the simplest and most widely used field test to verify batch-to-batch consistency of concrete and ensure that the fresh mix has not changed. In recent years, self-consolidating concrete (SCC) has been developed as an extremely high-workability mix that can flow and compact under its own weight without vibration. SCC is achieved by using advanced admixtures (superplasticizers) and viscosity-modifying agents to produce a fluid yet stable mix. It demonstrates that with appropriate mix design, one can obtain both high workability and high strength (through a low w/c ratio), which is valuable for placing concrete in structures with congested reinforcement or complex formwork. As the fresh concrete gradually sets and hardens, its properties transition to those of the hardened material. The density of normal-weight concrete is around 2400 kg/m³ (about 150 lb/ft³), and its compressive strength can range widely. Ordinary structural concrete typically achieves a 28-day compressive strength on the order of 20–40 MPa (3000–6000 psi), whereas high-strength concrete used in high-rise buildings or other special applications can reach strengths of 70 MPa (10,000 psi) or more. Concrete’s tensile strength is much lower than its compressive strength (on the order of only 10% of the compressive strength). For this reason, concrete is usually reinforced with steel bars (reinforcing steel, or “rebar”) to carry tension. The stress–strain curve of concrete in compression is nonlinear – unlike steel, concrete does not have a long elastic range or a well-defined yield point. It initially deforms approximately elastically at very low stress, but as stress increases, the curve gradually bends (due to microcracking and increasing inelasticity) and eventually reaches a peak, followed by a drop-off as the concrete fails by crushing. Because there is no sharply defined elastic limit, the modulus of elasticity of concrete is usually defined either as an initial tangent modulus or a secant modulus to a point on the curve (for example, to 50% of ultimate strength). For typical normal-strength concrete, a common empirical approximation is that the Young’s modulus is about 20–30 GPa (gigapascals), which corresponds to around 3 to 5 million psi. Higher-strength concretes have higher stiffness, though not in direct proportion to their strength. Volume stability is another important aspect of concrete performance. As concrete cures and dries, it undergoes shrinkage – a reduction in volume. Drying shrinkage occurs as water that is not chemically bound in hydration products gradually evaporates from the hardened concrete. If a large element (like a long slab or wall) is restrained while it is shrinking (for example, by the subgrade or by adjacent structural elements), tensile stresses can develop, potentially leading to cracking. To control crack formation, engineers often incorporate contraction joints (planned grooves or weak planes where shrinkage cracks can form in a controlled way) and provide sufficient distributed reinforcement (sometimes called temperature and shrinkage steel) to hold any cracks tightly closed. Thermal changes can also cause concrete to expand or contract. For instance, the heat released during cement hydration can raise the temperature of a large concrete mass in early-age, and subsequent cooling can induce tensile stresses. In hot weather, rapid surface drying and cooling of freshly placed concrete can cause a surface contraction while the interior is still warm, sometimes leading to a pattern of shallow cracks called thermal cracking. Measures such as gradual cooling of massive pours, insulating blankets, or using low-heat cements help mitigate these thermal stresses. In addition to shrinkage and thermal effects, concrete exhibits creep, which is the long-term gradual deformation under sustained load. Under a constant compressive stress (such as a heavy dead load on a column or beam), concrete will slowly deform over months and years beyond the initial elastic strain – this time-dependent strain is creep. Creep strain can be significant, sometimes on the same order as the initial elastic strain. It is influenced by factors like the concrete’s compressive strength, mix composition, age at loading, and the magnitude of the sustained stress (relative to strength). In structural design, creep is taken into account because it can increase deflections of beams over time and can cause redistribution of internal forces in statically indeterminate structures. However, keeping sustained stresses low (e.g. by using a higher concrete strength or adding extra reinforcement) can limit the magnitude of creep deformation. Durability is a critical aspect of civil materials engineering, and concrete must be designed to withstand the environmental conditions to which it will be exposed over decades. A key durability concern in cold climates is resistance to freeze–thaw cycles. When porous, saturated concrete freezes, the water in its pores expands about 9% in volume, which can create internal stresses that crack the concrete. Repeated freezing and thawing can cause progressive damage such as surface scaling and loss of strength. To improve freeze–thaw durability, modern concrete often includes air-entraining admixtures that create a network of microscopic air bubbles in the hardened concrete. These tiny air voids (~0.1 mm in size, with a proper spacing factor) act as pressure relief chambers for freezing water, thereby reducing internal stress. Air-entrained concrete can survive hundreds of freeze–thaw cycles without significant deterioration, whereas non-entrained concrete (especially if it has a high w/c ratio and hence high saturation) may suffer severe damage after just a few winters. Another long-term durability issue is chemical attack. For instance, sulfate attack occurs when sulfate ions (from sulfate-rich soils, groundwater, or seawater) penetrate the concrete and react with components of the cement paste (like calcium aluminate hydrates), forming expansive products such as ettringite and gypsum. These expansions can disrupt the concrete’s matrix, causing cracking and loss of cohesion over time. To combat sulfate attack, engineers can use sulfate-resistant cement (with low tricalcium aluminate content), incorporate pozzolanic materials, and ensure the concrete has a low permeability (achieved by a low w/c ratio and adequate curing). Similarly, a deleterious reaction called alkali–silica reaction (ASR) can occur between alkali hydroxides in cement paste and certain reactive silica minerals in some aggregates. ASR produces an expansive gel that swells as it absorbs water, leading to internal pressures and a pattern of map cracking in the concrete over years. To prevent ASR, engineers may use low-alkali cement and incorporate supplementary cementitious materials (like fly ash or slag) to tie up alkalis and dilute the reactive constituents, and they avoid using known reactive aggregates when possible. Moreover, many modern concrete mixtures include supplementary cementitious materials (SCMs) such as fly ash (a by-product from coal power plants), ground granulated blast-furnace slag, and silica fume. These SCMs participate in secondary pozzolanic reactions with the cement hydration products, refining the pore structure of the concrete and often improving its long-term performance. Concretes with SCMs tend to have lower permeability and higher later-age strength, which enhances durability by reducing the ingress of harmful agents. For example, adding fly ash or slag can significantly mitigate sulfate attack and alkali–silica reaction, while silica fume is well known for increasing strength and making the concrete matrix denser. The inclusion of SCMs also has sustainability benefits by recycling industrial by-products and reducing the required amount of Portland cement. Furthermore, reinforcement corrosion is a major threat to reinforced concrete structures, especially in harsh environments. Steel rebar embedded in concrete is normally protected by the high alkalinity of cement paste (which forms a thin passive oxide layer on the steel surface). However, if chloride ions (for example, from deicing salts or marine spray) penetrate through concrete and reach the steel, they can break down this passive layer and cause corrosion of the rebar. The corrosion products (iron oxides and hydroxides) occupy a greater volume than the original steel, exerting expansive pressure inside the concrete. This leads to cracking and spalling of the concrete cover, further accelerating deterioration. To improve durability against reinforcement corrosion, concrete can be made with low permeability (to slow chloride ingress) by using a low w/c ratio and SCMs; sufficient concrete cover thickness over the steel is provided to lengthen the travel path for chlorides; and sometimes corrosion-inhibiting admixtures or epoxy-coated (or galvanized) rebars are used. Good construction practices, such as adequate curing and avoiding any premature drying or thermal shock, also help produce a dense, impermeable concrete less vulnerable to these forms of attack. In summary, civil engineering materials like concrete must be designed not only for initial strength and workability but also for long-term durability. A well-proportioned concrete mix with an appropriate water–cement ratio, good-quality aggregates, and suitable admixtures can achieve the required strength while remaining sufficiently workable for placement. Through proper curing and quality control, the concrete will develop the desired mechanical properties (high compressive strength, adequate modulus, etc.) and a dense microstructure that resists deleterious agents. Attention to potential deterioration mechanisms – such as freeze–thaw damage, sulfate or chemical attack, ASR, and steel corrosion – is essential to ensure that structures made of concrete remain safe and serviceable throughout their intended life span. Ongoing advancements in materials engineering continue to improve concrete technology, yielding new types such as high-performance concrete (HPC) and fiber-reinforced concrete, which offer enhanced workability, strength, and durability for the infrastructure of the future.
1. “workability” 推荐翻译为 “和易性”,不应该译为“工作性” 2. “slump test” 推荐翻译为 “坍落度试验” 3. “water-cement ratio (w/c)” 推荐翻译为 “水灰比” 4. “curing” 推荐翻译为 “养护” 5. “hydration” 推荐翻译为 “水化作用” 6. “creep” 推荐翻译为 “徐变” 7. “shrinkage” 推荐翻译为 “收缩” 8. “segregation” 推荐翻译为 “离析” 9. “bleeding” 推荐翻译为 “泌水” 10. “air-entraining admixture” 推荐翻译为 “引气剂” 11. “freeze–thaw cycle” 推荐翻译为 “冻融循环” 12. “sulfate attack” 推荐翻译为 “硫酸盐侵蚀” 13. “alkali–silica reaction (ASR)” 推荐翻译为 “碱-硅反应” 14. “modulus of elasticity” 推荐翻译为 “弹性模量” 15. “map cracking” 推荐翻译为 “龟裂” 或 “龟甲裂缝”,不应该直译为“地图状裂缝”) 16. “thermal cracking” 推荐翻译为 “热裂缝” 或 “温度应力裂缝” 17. “rebar” 推荐翻译为 “钢筋” 18. “creep strain” 推荐翻译为 “徐变应变” 19. “impermeable concrete” / “dense concrete” 推荐翻译为 “低渗透性混凝土” 或 “致密混凝土” 20. “passive film” 推荐翻译为 “钝化膜” 21. “strength gain” 推荐翻译为 “强度发展” 或 “强度增长” 22. “superplasticizer” 推荐翻译为 “超塑化剂” 或 “高效减水剂”
18412
学术论文
自然科学
171
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Provisional Regulations for Transportation in Bond by Rail of Foreign Goods from Treaty Ports in South Manchuria to Treaty Ports in North Manchuria and China Proper via Mukden Kwanchengtze-Changchun and vice versa.(August, 1923, Draft) Article I. Applications for transportation in Bond by rail of foreign goods from the Coast Treaty Ports in South Manchuria and China Proper to the North Manchurian frontier stations of Manchuli and Suifenho and to Harbin via Kwanchengtze-Changchun and vice-versa, shall be made to the Customs by the carriers, i. e. (I) by the South Manchuria Railway Company for foreign goods entering South Manchuria at Antung, Dairen, Newchwang and Changchun, and (II) by the Chinese Eastern Railway for foreign goods entering North Manchuria at Manchouli, Suifenho and Kwanchengtze or despatched from Harbin and (III) by the Peking Mukden Railway for foreign goods entering South Manchuria at Newchwang and Shanhaikwan. A list of Treaty Ports and frontier stations to which this procedure applies is enclosed. The Customs notify the Railways on any subsequent changes therein. Article II. The South Manchuria Railway Company will enter into a guarantee with the Customs-either annual or for each consignment-that the goods accepted by it for transportation in Bond to North Manchuria are transported solely for the purpose of delivery to the Chinese Eastern Railway at Kwanchengtze-Changchun for further transmission in Bond by that Railway, and that those goods delivered to it at Kwanchentze-Changchun by the Chinese Eastern Railway for transmission in Bond are transported solely for the purpose of being transmitted to places in South Manchuria where Chinese Customs Houses are established, (it being understood that the responsibility of the South Manchuria Railway ceases directly the Goods have been taken over by the Chinese Eastern Railway at Kwangchentze-Changchun or delivered to consignees in South Manchuria after payment of duty), that such goods will only be conveyed in sealed carsor in open cars under seal, when necessary, and that in the event of any portion of the goods (northward-bound) accepted for transportation in Bond to Kwangchentze-Changchun for delivery to the Chinese Eastern Railway at that place for further transmission in Bond under seal not arriving within twenty days at Kwangchentze-Changchun, or (southward-bound) at destination in South Manchuria within the same time limit from the time the goods have been delivered to the South Manchuria Railway at Kwangzhentze-Changchun by the Chinese Eastern Railway, the South Manchuria Railway will pay to the Chinese Maritime Customs at port of entry a sum equal to twice the Import duty plus twice the Transit Dues leviable on that portion of the goods not arrived at Kwangchentze-Changchun (northward-bound) or at destination in South Manchuria (southward). If, owing to Customs or Railway holidays or unforeseen events, such as miscarriage of cargo, etc., other than those (force majeure) provided for separately hereafter, the South Manchuria Railway cannot delivery the goods (northward) at Kwangchentze-Changchun, or (southward) at destination in South Manchuria within the specified time limit, provided that satisfactory explanation be made to the Customs within seven days after expiry of such time limit, the Commissioner of Customs any defer enforcing the terms of the guarantee and extend the time limit for guarantee and extend the time limit for the recovery of the missing goods within a reasonable period. Accidents to the train caused by derailment, collision, fire, or other uncontrollable force (force majeure) will absolve the South Manchuria Railway from the fulfilment of the conditions as to the time limit within which the goods must reach Kwangchentze-Cahngchun or destination in South Manchuria, intact, provided that the Railway immediately informs the Customs as to the nature of the accident, takes proper precautions against pilfering, loss, etc., of the cargo, and detains the cars and/or the goods affected at the nearest convenient station in the event of the Customs seals being broken until receipt of the telegraphic instructions from the Customs Authorities as to what is to be further done. Article III. All cargo to be transported (northward) in Bond to Kwangchentze-Changchun for delivery to the Chinese Eastern Railway at that place for further transmission in Bond must be applied for to the Customs at port of entry on a special, “Transportation in Bond Application” in three sections marked “To Kwangchentze-Changchun for delivery to the Chinese Eastern Railway for transmission in Bond to….” Which will contain the following particulars, Marks, Numbers, Numbers of Packages, Description of Packing, Description of Cargo and Weight, Quantity, or Value. These details and to be filled in (according to the Bill of Lading or Train Bill which must accompany the application in duplicate) together with the number of the sealed cars conveying the goods, date, etc., on the triplicate sections of the application. After completion of all necessary Customs formalities at port of entry, the first Section of the application will be retained by the Customs at port of entry and the other two sections will ben forwarded to the Customs Officer at Kwanchengtze-Changchun, or to the Mukden Customs, if for transshipment to the Peking Mukden Railway, who will, after supervising transshipment of the cargo to cars of the Chinese Eastern Railway, or the Peking Mukden Railway, endorse these documents and forward them to the Commissioner of Customs at place of destination, who will return Section II, after endorsement, to the Customs at port of entry and will retain Section III for record. In the case of cargo in Bond from North Manchuria to South Manchuria delivered to the South Manchuria Railway by the Chinese Eastern Railway at Kwangchentze-Changchun Sections II and III of the “Transportation in Bond Application” will be endorsed by the South Manchuria Railway to the effect that the cargo in question has been accepted for transmission to South Manchuria in the terms of their annual guarantee. The South Manchuria Railway will undertake to deliver to the Customs at destination of the goods, such sealed covers containing Customs documents concerning the goods transported in Bond as may be handed to them by the Customs at port of entry or at Kwangchentze-Changchun. Article IV. Customs examination in the sense that packages are opened and content inspected need not necessarily take place at the port of entry in the case of cargo to be transported in Bond: but the right to subject any packages to such examination as he may think fit is reserved to the Commissioner of Customs at the port of entry. All such cargo, however, must be forwarded in cars sealed by the Customs at port of entry, or when there is not sufficient cargo to fill a car, in a sealed compartment to which access is impossible without tampering with the seals. Isolated packages of goods which it is desired to transport in Bond may be corded and sealed. The question of providing seals and sealing appliances will be arranged for by the Railway Authorities in consultation with the Customs. Article V. On arrival at Kwangchentze-Changchun ori at Mukden the seals will be inspected by the Customs Officer, and if found to be intact permission will be granted to transship such cargo into the Chinese Eastern Railway cars or Peking Mukden Railway cars as the case may be, under Customs supervision (or to be placed in registered godowns until transshipment into the Chinese Eastern Railway cars or Peking Mukden Railway car is ready to take place.) All cargo tallied out and again loaded into the new cars at Kwangchentze-Changchun or at Mukden must agree with the particulars contained in Sections II and III of the original “Transportation in Bond Application” and the train Bill or Bill of Lading accompanying the goods. After all cargo has been duly transhipped, the Customs Officer will seal the cars or the specially reserved portion of cars and will allow them to proceed to destination. At Kwangchentze-Changchun or at Mukden no one but duly authorized Chinese Customs Officer shall be permitted to break or remove Customs seals from Railway cars. The unsealing and resealing of cars, etc., is to be entirely in hands of the Customs Offices. Article VI. If at any stage of the journey from port of entry to Kwangchentze-Changchun or vice-versa from Kwangchentze-Changchun to destination in South Manchuria, the Customs selas are discovered to be damaged or missing thus rendering access to, and tampering with, bonded cargo possible; or, of a sealed car and/or bonded goods be damaged en route either through accident as described in Article II or through any other cause; the Railway will immediately communicate by railway telegraph with the nearest convenient station pending receipt of telegraphic instructions from the Customs. It rests then with the Customs Authorities to appoint a special employee or employees – whom the Railway will furnish with free passes for travelling on their lines, - to carry out investigations into the causes of damage to, or loss of, seals, or to commission the Railway to make such investigations, or to authorize the dispatch of the car and/or the goods to one of the Customs stations for investigation. The result of the investigation will be duly entered in a Protocol specially kept for this purpose and signed by the Customs and Railway employees concerned. If at the investigation or upon inspection and comparison of the cargo with documents at the place of accident or at the Customs station to which the car and/or the goods have been directed to proceed, the cargo be found in full agreement with the documents, it will be resealed by the Customs and allowed to proceed without hindrance at the first opportunity, and the necessary remarks and corrections will be made in the accompanying documents. Should however, the loss of the Customs seals be not properly accounted for, and the Customs have cause to suspect tampering with, or substitution of goods, the terms of the guarantee will be enforced by the Customs. If in case of damage through force majeure, it is intended to sell the remaining or the damaged cargo on the spot, Customs permission must be obtained before hand. Import duty (and Transit Dues if liviable) must be paid by the purchaser before delivery of the sold goods. Article VII. The transportation in Bond regulations do not apply to native goods for the present, munitions of war, explosives and other dangerous goods, nor to hand baggage and/or registered luggage which are subject to usual examination and duty treatment at port of entry as at present exercised. Article VIII. The above regulations are provisional, and subject to alteration, addition, or cancellation as experience may prove to be necessary.
Provisional Regulations for Transportation in Bond by Rail of Foreign Goods from Treaty Ports in South Manchuria to Treaty Ports in North Manchuria and China Proper via Mukden Kwanchengtze-Changchun and vice versa.(August, 1923, Draft) Article I. Applications for transportation in Bond by rail of foreign goods from the Coast Treaty Ports in South Manchuria and China Proper to the North Manchurian frontier stations of Manchuli and Suifenho and to Harbin via Kwanchengtze-Changchun and vice-versa, shall be made to the Customs by the carriers, i. e. (I) by the South Manchuria Railway Company for foreign goods entering South Manchuria at Antung, Dairen, Newchwang and Changchun, and (II) by the Chinese Eastern Railway for foreign goods entering North Manchuria at Manchouli, Suifenho and Kwanchengtze or despatched from Harbin and (III) by the Peking Mukden Railway for foreign goods entering South Manchuria at Newchwang and Shanhaikwan. A list of Treaty Ports and frontier stations to which this procedure applies is enclosed. The Customs notify the Railways on any subsequent changes therein. Article II. The South Manchuria Railway Company will enter into a guarantee with the Customs-either annual or for each consignment-that the goods accepted by it for transportation in Bond to North Manchuria are transported solely for the purpose of delivery to the Chinese Eastern Railway at Kwanchengtze-Changchun for further transmission in Bond by that Railway, and that those goods delivered to it at Kwanchentze-Changchun by the Chinese Eastern Railway for transmission in Bond are transported solely for the purpose of being transmitted to places in South Manchuria where Chinese Customs Houses are established, (it being understood that the responsibility of the South Manchuria Railway ceases directly the Goods have been taken over by the Chinese Eastern Railway at Kwangchentze-Changchun or delivered to consignees in South Manchuria after payment of duty), that such goods will only be conveyed in sealed carsor in open cars under seal, when necessary, and that in the event of any portion of the goods (northward-bound) accepted for transportation in Bond to Kwangchentze-Changchun for delivery to the Chinese Eastern Railway at that place for further transmission in Bond under seal not arriving within twenty days at Kwangchentze-Changchun, or (southward-bound) at destination in South Manchuria within the same time limit from the time the goods have been delivered to the South Manchuria Railway at Kwangzhentze-Changchun by the Chinese Eastern Railway, the South Manchuria Railway will pay to the Chinese Maritime Customs at port of entry a sum equal to twice the Import duty plus twice the Transit Dues leviable on that portion of the goods not arrived at Kwangchentze-Changchun (northward-bound) or at destination in South Manchuria (southward). If, owing to Customs or Railway holidays or unforeseen events, such as miscarriage of cargo, etc., other than those (force majeure) provided for separately hereafter, the South Manchuria Railway cannot delivery the goods (northward) at Kwangchentze-Changchun, or (southward) at destination in South Manchuria within the specified time limit, provided that satisfactory explanation be made to the Customs within seven days after expiry of such time limit, the Commissioner of Customs any defer enforcing the terms of the guarantee and extend the time limit for guarantee and extend the time limit for the recovery of the missing goods within a reasonable period. Accidents to the train caused by derailment, collision, fire, or other uncontrollable force (force majeure) will absolve the South Manchuria Railway from the fulfilment of the conditions as to the time limit within which the goods must reach Kwangchentze-Cahngchun or destination in South Manchuria, intact, provided that the Railway immediately informs the Customs as to the nature of the accident, takes proper precautions against pilfering, loss, etc., of the cargo, and detains the cars and/or the goods affected at the nearest convenient station in the event of the Customs seals being broken until receipt of the telegraphic instructions from the Customs Authorities as to what is to be further done. Article III. All cargo to be transported (northward) in Bond to Kwangchentze-Changchun for delivery to the Chinese Eastern Railway at that place for further transmission in Bond must be applied for to the Customs at port of entry on a special, “Transportation in Bond Application” in three sections marked “To Kwangchentze-Changchun for delivery to the Chinese Eastern Railway for transmission in Bond to….” Which will contain the following particulars, Marks, Numbers, Numbers of Packages, Description of Packing, Description of Cargo and Weight, Quantity, or Value. These details and to be filled in (according to the Bill of Lading or Train Bill which must accompany the application in duplicate) together with the number of the sealed cars conveying the goods, date, etc., on the triplicate sections of the application. After completion of all necessary Customs formalities at port of entry, the first Section of the application will be retained by the Customs at port of entry and the other two sections will ben forwarded to the Customs Officer at Kwanchengtze-Changchun, or to the Mukden Customs, if for transshipment to the Peking Mukden Railway, who will, after supervising transshipment of the cargo to cars of the Chinese Eastern Railway, or the Peking Mukden Railway, endorse these documents and forward them to the Commissioner of Customs at place of destination, who will return Section II, after endorsement, to the Customs at port of entry and will retain Section III for record. In the case of cargo in Bond from North Manchuria to South Manchuria delivered to the South Manchuria Railway by the Chinese Eastern Railway at Kwangchentze-Changchun Sections II and III of the “Transportation in Bond Application” will be endorsed by the South Manchuria Railway to the effect that the cargo in question has been accepted for transmission to South Manchuria in the terms of their annual guarantee. The South Manchuria Railway will undertake to deliver to the Customs at destination of the goods, such sealed covers containing Customs documents concerning the goods transported in Bond as may be handed to them by the Customs at port of entry or at Kwangchentze-Changchun. Article IV. Customs examination in the sense that packages are opened and content inspected need not necessarily take place at the port of entry in the case of cargo to be transported in Bond: but the right to subject any packages to such examination as he may think fit is reserved to the Commissioner of Customs at the port of entry. All such cargo, however, must be forwarded in cars sealed by the Customs at port of entry, or when there is not sufficient cargo to fill a car, in a sealed compartment to which access is impossible without tampering with the seals. Isolated packages of goods which it is desired to transport in Bond may be corded and sealed. The question of providing seals and sealing appliances will be arranged for by the Railway Authorities in consultation with the Customs. Article V. On arrival at Kwangchentze-Changchun ori at Mukden the seals will be inspected by the Customs Officer, and if found to be intact permission will be granted to transship such cargo into the Chinese Eastern Railway cars or Peking Mukden Railway cars as the case may be, under Customs supervision (or to be placed in registered godowns until transshipment into the Chinese Eastern Railway cars or Peking Mukden Railway car is ready to take place.) All cargo tallied out and again loaded into the new cars at Kwangchentze-Changchun or at Mukden must agree with the particulars contained in Sections II and III of the original “Transportation in Bond Application” and the train Bill or Bill of Lading accompanying the goods. After all cargo has been duly transhipped, the Customs Officer will seal the cars or the specially reserved portion of cars and will allow them to proceed to destination. At Kwangchentze-Changchun or at Mukden no one but duly authorized Chinese Customs Officer shall be permitted to break or remove Customs seals from Railway cars. The unsealing and resealing of cars, etc., is to be entirely in hands of the Customs Offices. Article VI. If at any stage of the journey from port of entry to Kwangchentze-Changchun or vice-versa from Kwangchentze-Changchun to destination in South Manchuria, the Customs selas are discovered to be damaged or missing thus rendering access to, and tampering with, bonded cargo possible; or, of a sealed car and/or bonded goods be damaged en route either through accident as described in Article II or through any other cause; the Railway will immediately communicate by railway telegraph with the nearest convenient station pending receipt of telegraphic instructions from the Customs. It rests then with the Customs Authorities to appoint a special employee or employees – whom the Railway will furnish with free passes for travelling on their lines, - to carry out investigations into the causes of damage to, or loss of, seals, or to commission the Railway to make such investigations, or to authorize the dispatch of the car and/or the goods to one of the Customs stations for investigation. The result of the investigation will be duly entered in a Protocol specially kept for this purpose and signed by the Customs and Railway employees concerned. If at the investigation or upon inspection and comparison of the cargo with documents at the place of accident or at the Customs station to which the car and/or the goods have been directed to proceed, the cargo be found in full agreement with the documents, it will be resealed by the Customs and allowed to proceed without hindrance at the first opportunity, and the necessary remarks and corrections will be made in the accompanying documents. Should however, the loss of the Customs seals be not properly accounted for, and the Customs have cause to suspect tampering with, or substitution of goods, the terms of the guarantee will be enforced by the Customs. If in case of damage through force majeure, it is intended to sell the remaining or the damaged cargo on the spot, Customs permission must be obtained before hand. Import duty (and Transit Dues if liviable) must be paid by the purchaser before delivery of the sold goods. Article VII. The transportation in Bond regulations do not apply to native goods for the present, munitions of war, explosives and other dangerous goods, nor to hand baggage and/or registered luggage which are subject to usual examination and duty treatment at port of entry as at present exercised. Article VIII. The above regulations are provisional, and subject to alteration, addition, or cancellation as experience may prove to be necessary.
考点1:“Transportation in Bond”应当译为“保税运输”,或者按照中国海关当时的叫法“通车运”,此时这一短语可以调整为动词使用 考点 2 :“Newchwang”应当译为“营口” 考点 3 :“guarantee”应当译为“保结” 考点 4 :“Chinese Customs Houses”应当译为“中国海关” 考点 5 :“sealed carsor”应当译为“封固车厢”或者和“封制车厢” 考点 6 :“Transit Dues”应当译为“子口税” 考点 7 :“Commissioner of Customs”应当译为“海关税务司” 考点 8 :“Train Bill”应当译为“铁路运单”或者“车单” 考点 9 :“annual guarantee”应当译为“常年保结” 考点 10 :“Provisional Regulations”应当译为“暂行章程” 考点 11 :“registered godowns”应当译为“关栈”
19859
垂类场景
法律
163
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Neil Diamond to Receive Inspirational Lifetime Achievement Award at 2025 Carousel Ball, Goo Goo Dolls to Perform The charity gala, benefiting Children's Diabetes Foundation, will be held in October. UPDATE (July 9): Goo Goo Dolls are set to perform at the 39th annual Carousel Ball at the Hyatt Regency Denver on Oct. 11, honoring music icon Neil Diamond. Previous stars who’ve headlined the ball included Sir Elton John, Celine Dion, Whitney Houston and John Legend. Goo Goo Dolls have landed two No. 1 hits on Billboard’s all-genre Radio Songs chart – “Iris,” which held the top spot for a then record 18 weeks in 1998, and “Slide.” “We’re excited to take the stage at The Carousel Ball and support an amazing cause,” said Goo Goo Dolls bandmates John Rzeznik and Robby Takac in a joint statement. “The progress the Children’s Diabetes Foundation has made in diabetes research and patient care has brought new hope to countless families and we’re honored to be a part of their special night.” PREVIOUSLY (April 29): Neil Diamond is set to receive the Inspirational Lifetime Achievement Award by the Children’s Diabetes Foundation at the 39th Annual Carousel Ball on Oct. 11. The event will be held at the Hyatt Regency Hotel in Denver, near the home of CDF’s primary operations and its clinic and research facility, the Barbara Davis Center for Diabetes. Diamond is just the third recipient of the award, following Sidney Poitier (2016) and Diane Warren (2024). Proceeds from The Carousel Ball benefit CDF and focus on patient support, awareness and diabetes research. The Carousel Balls, which date to 1978, collectively have raised more than $117 million. Diamond, 84, has served on CDF’s advisory board and has attended the organization’s fundraisers for many years, including performing in Denver in 2001, as well as at The Carousel of Hope Ball in Beverly Hills in 2012, where he memorably sang an impromptu duet of “Sweet Caroline” with George Clooney. “It’s been my absolute joy to have supported the Children’s Diabetes Foundation for so many wonderful years,” Diamond said in a statement. “Barbara Davis and her organization have done an immense amount of good helping kids, adults and their families facing a difficult diagnosis.” “Neil Diamond is a once-in-a-generation talent,” event chair Dana Davis and honorary chair Barbara Davis said in a joint statement. “As a bestselling recording artist the world over, he is an undeniable force whose songwriting has earned him countless musical accolades. … For over 35 years, Neil has lent his timeless voice to our efforts to find a cure for diabetes. On behalf of the patients and researchers his generosity has touched, we can’t think of anyone more deserving of our Inspirational Lifetime Achievement Award.” Diamond’s career spans nearly 60 years. He landed his first top 10 hit on the Billboard Hot 100, “Cherry, Cherry,” in October 1966. He has notched three No. 1 hits on the Hot 100 — “Cracklin’ Rosie” (1970), “Song Sung Blue” (1972) and “You Don’t Bring Me Flowers,” a duet with Barbra Streisand (1978). Diamond finally topped the Billboard 200 in 2008, nearly 42 years after he first made that chart, with Home Before Dark. Diamond was elected to the Songwriters Hall of Fame in 1984 and has received two additional awards from that organization — the Sammy Cahn Lifetime Achievement in 2000 and the Johnny Mercer Award (their top honor) in 2018. Additionally, Diamond received MusiCares’ Person of the Year Award in 2009, the Kennedy Center Honors in 2011 and a lifetime achievement award from the Recording Academy in 2018. He was voted into the Rock & Roll Hall of Fame in 2011. His achievements also include a Grammy, a Golden Globe Award, an American Music Award, an ASCAP Film and Television Award and a Billboard Icon Award. In addition to the award to Diamond, The MacMillan Family will be honored with the High Hopes Tribute Award for their nearly three decades of commitment as supporters of CDF. To purchase tickets and tables and learn more about the event, visit the Children’s Diabetes Foundation’s site. Jelly Roll, Lainey Wilson, Luke Combs & Rissi Palmer Among ACM Special Awards Honorees This year's ACM Honors will be held at Nashville venue The Pinnacle on Aug. 20. Jelly Roll, Lainey Wilson, Eric Church, Rissi Palmer, Randy Travis, Luke Combs and Cody Johnson are among this year’s recipients of the Academy of Country Music’s special awards, and will be feted during the 18th Academy of Country Music Honors celebration, which will take place Wednesday, Aug. 20, at The Pinnacle in Nashville. Jelly Roll will be honored with the ACM lifting lives award, which recognizes a country music artist, duo/group or industry member who has devoted themselves to improving lives through music and is committed to serving others. The honor is voted on by the ACM lifting lives board of directors. Jelly Roll is being honored for his philanthropic work. He teamed with Live Nation to donate $1 from each ticket sold on his Backroad Baptism Tour, raising more than $600,000 for at-risk youth. He has also visited over 10 juvenile facilities, four rehabilitation centers and several homeless shelters to help bring hope and encouragement. He also testified before Congress to advocate for the Fentanyl Eradication and Narcotics Deterrence Off Fentanyl Act. Church and Ben Vaughn, president/CEO of Warner Chappell Music Nashville until his passing earlier this year, are the recipients of the ACM icon award, given to presented to a country music artist, duo/group or industry leader who has advanced the popularity of the genre through their contributions throughout their career. Church, a seven-time ACM Award winner, spearheaded the Concert for Carolina alongside Luke Combs, raising more than $24 million for Hurricane Helene relief efforts. For Church’s 2024 release “Darkest Hour,” he signed over all of his publishing royalties to aid those impacted by the floods in his homestate of North Carolina. During his career leading Warner Chappell Music Nashville, Vaughn was instrumental in developing numerous artists and songwriters including Dan+Shay, Kacey Musgraves, Lady A, Thomas Rhett and Chris Stapleton. Johnson is the recipient of the ACM spirit award, inspired by the late country music legend Merle Haggard. The award is presented to a singer-songwriter who continues the legacy of Haggard by following their own path, crafting compelling songs and presenting them through high-caliber performances. Johnson has earned two No. 1 Billboard Country Airplay hits, “‘Til You Can’t” and “The Painter.” Combs will be honored with the ACM international award, which celebrates a country music artist, duo/group or industry leader for outstanding contributions to the growth of country music globally. Combs’ headlining tours have sold out in the U.K., Ireland, the Netherlands and Australia, demonstrating his global appeal and fanbase. Palmer will be feted with the ACM’s lift every voice award, which is presented to a country music artist, duo/group, industry leader or affiliate/partner who elevates underrepresented voices in the country music genre. The nominee for this category is proposed by the rising leaders in the ACM’s LEVel UP: Lift Every Voice professional development program. In addition to leading her own music career, Palmer is the creator and host of Apple Music’s Color Me Country Radio with Rissi Palmer, which celebrates diversity and inclusion in the country music genre and spotlights artists of color. Mac McAnally will be recognized with the ACM poet’s award, which honors a country music songwriter for longstanding musical contributions to country music. In addition to writing his own music, this Nashville Songwriters Hall of Fame member has had songs recorded by Jimmy Buffett (“It’s My Job”) Kenny Chesney (“Down the Road”), Alabama (“Old Flame”), Sawyer Brown (“I Will Leave the Light On”), Shenandoah (“Two Dozen Roses”) and more. Country Music Hall of Fame member Travis will be honored as the recipient of the ACM milestone award, which is presented to a country music artist, duo/group or industry leader for outstanding achievements in country music during the preceding calendar year. Travis incorporated AI-assisted vocals to re-imagine and revive older and previously unheard songs, preserving and continuing his unmistakable sound. The ACM film award, which recognizes an outstanding movie, series or feature film released during the preceding calendar year that prominently features country music, will be presented for the film Twisters. The film’s soundtrack featured music from Luke Combs, Megan Moroney, Jelly Roll and more. Executive Lori Badgett is the recipient of the ACM service award, which is given to an outstanding country music artist, duo/group or industry leader to recognize years of dedication and service to the ACM. In her work at City National Bank, Badgett has helped establish the bank’s presence in Nashville and she now plays an essential role in the ACM and its charitable arm ACM Lifting Lives, having chaired both organizations. Badgett’s leadership was instrumental in the Academy’s transition from Los Angeles to Nashville and she continues to serve on multiple committees for both organizations.
Neil Diamond to Receive Inspirational Lifetime Achievement Award at 2025 Carousel Ball, Goo Goo Dolls to Perform The charity gala, benefiting Children's Diabetes Foundation, will be held in October. UPDATE (July 9): Goo Goo Dolls are set to perform at the 39th annual Carousel Ball at the Hyatt Regency Denver on Oct. 11, honoring music icon Neil Diamond. Previous stars who’ve headlined the ball included Sir Elton John, Celine Dion, Whitney Houston and John Legend. Goo Goo Dolls have landed two No. 1 hits on Billboard’s all-genre Radio Songs chart – “Iris,” which held the top spot for a then record 18 weeks in 1998, and “Slide.” “We’re excited to take the stage at The Carousel Ball and support an amazing cause,” said Goo Goo Dolls bandmates John Rzeznik and Robby Takac in a joint statement. “The progress the Children’s Diabetes Foundation has made in diabetes research and patient care has brought new hope to countless families and we’re honored to be a part of their special night.” PREVIOUSLY (April 29): Neil Diamond is set to receive the Inspirational Lifetime Achievement Award by the Children’s Diabetes Foundation at the 39th Annual Carousel Ball on Oct. 11. The event will be held at the Hyatt Regency Hotel in Denver, near the home of CDF’s primary operations and its clinic and research facility, the Barbara Davis Center for Diabetes. Diamond is just the third recipient of the award, following Sidney Poitier (2016) and Diane Warren (2024). Proceeds from The Carousel Ball benefit CDF and focus on patient support, awareness and diabetes research. The Carousel Balls, which date to 1978, collectively have raised more than $117 million. Diamond, 84, has served on CDF’s advisory board and has attended the organization’s fundraisers for many years, including performing in Denver in 2001, as well as at The Carousel of Hope Ball in Beverly Hills in 2012, where he memorably sang an impromptu duet of “Sweet Caroline” with George Clooney. “It’s been my absolute joy to have supported the Children’s Diabetes Foundation for so many wonderful years,” Diamond said in a statement. “Barbara Davis and her organization have done an immense amount of good helping kids, adults and their families facing a difficult diagnosis.” “Neil Diamond is a once-in-a-generation talent,” event chair Dana Davis and honorary chair Barbara Davis said in a joint statement. “As a bestselling recording artist the world over, he is an undeniable force whose songwriting has earned him countless musical accolades. … For over 35 years, Neil has lent his timeless voice to our efforts to find a cure for diabetes. On behalf of the patients and researchers his generosity has touched, we can’t think of anyone more deserving of our Inspirational Lifetime Achievement Award.” Diamond’s career spans nearly 60 years. He landed his first top 10 hit on the Billboard Hot 100, “Cherry, Cherry,” in October 1966. He has notched three No. 1 hits on the Hot 100 — “Cracklin’ Rosie” (1970), “Song Sung Blue” (1972) and “You Don’t Bring Me Flowers,” a duet with Barbra Streisand (1978). Diamond finally topped the Billboard 200 in 2008, nearly 42 years after he first made that chart, with Home Before Dark. Diamond was elected to the Songwriters Hall of Fame in 1984 and has received two additional awards from that organization — the Sammy Cahn Lifetime Achievement in 2000 and the Johnny Mercer Award (their top honor) in 2018. Additionally, Diamond received MusiCares’ Person of the Year Award in 2009, the Kennedy Center Honors in 2011 and a lifetime achievement award from the Recording Academy in 2018. He was voted into the Rock & Roll Hall of Fame in 2011. His achievements also include a Grammy, a Golden Globe Award, an American Music Award, an ASCAP Film and Television Award and a Billboard Icon Award. In addition to the award to Diamond, The MacMillan Family will be honored with the High Hopes Tribute Award for their nearly three decades of commitment as supporters of CDF. To purchase tickets and tables and learn more about the event, visit the Children’s Diabetes Foundation’s site. Jelly Roll, Lainey Wilson, Luke Combs & Rissi Palmer Among ACM Special Awards Honorees This year's ACM Honors will be held at Nashville venue The Pinnacle on Aug. 20. Jelly Roll, Lainey Wilson, Eric Church, Rissi Palmer, Randy Travis, Luke Combs and Cody Johnson are among this year’s recipients of the Academy of Country Music’s special awards, and will be feted during the 18th Academy of Country Music Honors celebration, which will take place Wednesday, Aug. 20, at The Pinnacle in Nashville. Jelly Roll will be honored with the ACM lifting lives award, which recognizes a country music artist, duo/group or industry member who has devoted themselves to improving lives through music and is committed to serving others. The honor is voted on by the ACM lifting lives board of directors. Jelly Roll is being honored for his philanthropic work. He teamed with Live Nation to donate $1 from each ticket sold on his Backroad Baptism Tour, raising more than $600,000 for at-risk youth. He has also visited over 10 juvenile facilities, four rehabilitation centers and several homeless shelters to help bring hope and encouragement. He also testified before Congress to advocate for the Fentanyl Eradication and Narcotics Deterrence Off Fentanyl Act. Church and Ben Vaughn, president/CEO of Warner Chappell Music Nashville until his passing earlier this year, are the recipients of the ACM icon award, given to presented to a country music artist, duo/group or industry leader who has advanced the popularity of the genre through their contributions throughout their career. Church, a seven-time ACM Award winner, spearheaded the Concert for Carolina alongside Luke Combs, raising more than $24 million for Hurricane Helene relief efforts. For Church’s 2024 release “Darkest Hour,” he signed over all of his publishing royalties to aid those impacted by the floods in his homestate of North Carolina. During his career leading Warner Chappell Music Nashville, Vaughn was instrumental in developing numerous artists and songwriters including Dan+Shay, Kacey Musgraves, Lady A, Thomas Rhett and Chris Stapleton. Johnson is the recipient of the ACM spirit award, inspired by the late country music legend Merle Haggard. The award is presented to a singer-songwriter who continues the legacy of Haggard by following their own path, crafting compelling songs and presenting them through high-caliber performances. Johnson has earned two No. 1 Billboard Country Airplay hits, “‘Til You Can’t” and “The Painter.” Combs will be honored with the ACM international award, which celebrates a country music artist, duo/group or industry leader for outstanding contributions to the growth of country music globally. Combs’ headlining tours have sold out in the U.K., Ireland, the Netherlands and Australia, demonstrating his global appeal and fanbase. Palmer will be feted with the ACM’s lift every voice award, which is presented to a country music artist, duo/group, industry leader or affiliate/partner who elevates underrepresented voices in the country music genre. The nominee for this category is proposed by the rising leaders in the ACM’s LEVel UP: Lift Every Voice professional development program. In addition to leading her own music career, Palmer is the creator and host of Apple Music’s Color Me Country Radio with Rissi Palmer, which celebrates diversity and inclusion in the country music genre and spotlights artists of color. Mac McAnally will be recognized with the ACM poet’s award, which honors a country music songwriter for longstanding musical contributions to country music. In addition to writing his own music, this Nashville Songwriters Hall of Fame member has had songs recorded by Jimmy Buffett (“It’s My Job”) Kenny Chesney (“Down the Road”), Alabama (“Old Flame”), Sawyer Brown (“I Will Leave the Light On”), Shenandoah (“Two Dozen Roses”) and more. Country Music Hall of Fame member Travis will be honored as the recipient of the ACM milestone award, which is presented to a country music artist, duo/group or industry leader for outstanding achievements in country music during the preceding calendar year. Travis incorporated AI-assisted vocals to re-imagine and revive older and previously unheard songs, preserving and continuing his unmistakable sound. The ACM film award, which recognizes an outstanding movie, series or feature film released during the preceding calendar year that prominently features country music, will be presented for the film Twisters. The film’s soundtrack featured music from Luke Combs, Megan Moroney, Jelly Roll and more. Executive Lori Badgett is the recipient of the ACM service award, which is given to an outstanding country music artist, duo/group or industry leader to recognize years of dedication and service to the ACM. In her work at City National Bank, Badgett has helped establish the bank’s presence in Nashville and she now plays an essential role in the ACM and its charitable arm ACM Lifting Lives, having chaired both organizations. Badgett’s leadership was instrumental in the Academy’s transition from Los Angeles to Nashville and she continues to serve on multiple committees for both organizations.
考点1:【Carousel Ball】应统一译为“旋转木马慈善舞会” 考点2:【Inspirational Lifetime Achievement Award】推荐译为“励志终身成就奖” 考点3:【Goo Goo Dolls】为知名摇滚乐队,应保留原文名,不加译名,若解释性加注,可注明为“咕咕玩偶(Goo Goo Dolls)”。 考点4:【Songwriters Hall of Fame】应统一译为“词曲作者名人堂”。 考点5:【MusiCares' Person of the Year】应译为“年度音乐关怀人物奖” 考点6:【AI-assisted vocals】推荐译为“AI辅助的人声重建技术”,不可译为“AI歌声”或“机器人声音”。 考点7:奖项名称应统一保留机构缩写,如:“ASCAP Film and Television Award”→ “ASCAP影视奖”;“ACM Spirit Award”→ “ACM精神奖” 考点8:【ACM Lift Every Voice Award】应译为“ACM群声之上奖”
1c0e1
新闻资讯
新闻报道
71
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: This then returns me to the question of an aesthetics of care. If as I have already argued a care based ethics helps raise questions of justice and ‘other regarding effort’, how might art making be judged from this perspective? The starting point is the notion of relations and the simplistic statement that art making takes place in a series of relational acts, some more explicit and intentional than others. Where an ethics of care focuses upon the values inherent, exhibited or perhaps desired within these human interdependencies, the aesthetics of care seeks to focus upon how the sensory and affective are realised in human relations fostered in art projects. The French art theorist Nicolas Bourriaud is a useful point of departure here in his work on relational aesthetics. Bourriaud defines a relational aesthetic as a ‘set of artistic practices which take as their theoretical and practical point of departure the whole of human relations and their social context, rather than an independent and private space’ (2011, 113) and his book announced from the perspective of the late 1990s French visual art scene how ‘for some years now, there has been an upsurge of convivial, user-friendly artistic projects, festive, collective and participatory, exploring the varied potential in the relationship to the other’ (61). Bourriaud’s work appears to attach an implicit value to this upsurge, with the ‘angelic programme’ I mentioned at the top of the article being instigated in order ‘to patiently re-stitch the relational fabric’ (36) and ‘turn the beholder into the neighbour’ (43). However, ultimately he is more concerned with the formal aspects of this trajectory than the potential that new relational practices have for announcing or creating a fairer world. Inventing new neighbourly relations today, he asserts, is disconnected from programmes that seek to foster ‘happier tomorrows’ (45). As the writing above on the ethics of care would suggest, and the contract of mutual regard urged by Geras, the values that can be materialised in the convivial should in fact be the ground on which happier tomorrows are built. The power of the concept of relational aesthetics is weakened by the fact it does not suggest why relations with others might be endorsed or what type of relations we might aspire to develop. If the socially critical inference in the notion of relational aesthetics is to have greater explicit ethical weight, and to move from a moderately interesting description of a form to a movement with more normative clout, it needs to be refigured. And the argument here is that thinking in terms of an aesthetics of care might provide this reorientation. An ‘aesthetics of care’ is then about a set of values realised in a relational process that emphasise engagements between individuals or groups over time. It is one that might consist of small creative encounters or large-scale exhibitions, but it is always one that notices inter-human relations in both the creation and the display of art projects. It is an aesthetics that is unafraid to lay bare what Shannon Jackson calls the ‘supporting infrastructures of [ … ] living beings’ (2011, 39), but importantly this is an aesthetics that could both present those mutually beneficial structures and foster them. It would not pretend to a distinction between a process and an outcome because both might stimulate affective solidarity between people – perhaps participant to participant or performer to audience. There is a sense that this aesthetics would value intimacy, but it would not be at the expense of what Nato Thompson refers to as ‘explicitly local, long-term, and community-based’ engagement (2012, 31). While care might be exhibited fleetingly, it is more likely that care aesthetics would be realised in more enduring, crafted encounters between people. Seeking to overcome widespread social indifference implies commitment to deep and extended processes. At the beginning of the article it was noted that the ethics of care is a reference to both a set of values and a practice. This is repeated in this proposal for an aesthetics of care, so that it suggests both a demonstration of mutual regard, but simultaneously it instigates a process that is seeking to create or secure it. Amin asserts that there is a process of cultivation in his project for overcoming the disregard experienced in a society of strangers. It is a ‘craft that requires continual attentiveness and care, such that empathy – for objects, projects, nature, the commons – can spread as a public sentiment that also serves to regulate feelings among strangers’ (2012, 7). Attentiveness (a term also found in Tronto (2013, 34)) is both at the heart of the creative process, and the outcome of it. An aesthetics of care is, therefore, a sensory ethical practice, that following Robinson, involves ‘not only learning how to be attentive and patient, how to listen and respond, but also how to rethink our own attitudes about difference and exclusion’ (1999, 164). The difficulty in nurturing an ethics of care through an aesthetic process should of course be acknowledged. Kester in talking about participatory modes of art making explains it as a ‘temporally extensive form of social interaction in which models of expression, enunciation, and reception are continuously modified and reciprocally responsive’ (2011, 112). It is a form of crafted caring where learning to create, respond and be in close dialogue with others is vital for the quality of the experience: but it is a ‘temporally extensive form’ because it needs to be ‘continuously modified’ as it is practiced. In order to outline something of the shape of that form, the remainder of this article will divide care aesthetics into preparation, execution and exhibition, where each of these moments might be minutely connected but they are also likely to be dispersed over a long period. First, preparation would involve an openness and honesty of intention, the selection of artists or participants and questions of the location of a project. Decisions about accessibility (whether in terms of the appropriateness of the space for disabled people, the location in terms of costs of travel or the timing for people with different commitments) are not mundane organisational matters, but crucial ethical propositions. In being taken in reference to the ethics of care, they will imbue the project with an affective sense of the importance of mutual respect and regard. Jackson’s ‘supporting infrastructures’ are not the hidden mechanism of creative endeavour but a valued component of the aesthetics. Preparation is, therefore, paradoxically part of the exhibition within this mode of artistic project: it can demonstrate and model a form of mutual regard. There is a sensory quality in the relationships to which a project that prepares in this way aspires. The notion of execution focuses on the process of collaborative working on artistic projects which forge inter-human relationships. The emerging connections between individuals coalescing in this process have an aesthetics – a shape, feel, sensation and affect. This does not exist within one particular person or object of the work, but appears in-between those involved, so that there is a sensory quality of the process and outcome that cannot be disaggregated from the collective effort. This connects to Richard Sennett’s work on the history and practice of cooperation. His conviction that a practice of working together, and his central example of the Hull House settlement in Chicago, demonstrates that a shared commitment to building caring relations turns ‘people outward in shared, symbolic acts’ and these in turn have a potential place in countering a society that is figured ‘brutally simple: us-againstthem coupled with you-are-on-your-own’ (Sennett 2012, 280). For theatre this form of cooperation might involve a challenge to the quality and texture of a rehearsal or devising experience so that the reciprocity of gradual creation is valued over and above the discipline of the single-minded voyage towards the first night. Those intimate negotiations are the aesthetics of the project, and not merely an unremarkable preparatory period. Borrowing from David Gauntlett’s work on craft, there is a suggestion that within a creative process the power is realised ‘in gentle and quiet ways, with no need for grand celebratory announcements’ (2011, 66). This is not meant as a dismissal of the art of public theatre making as part of the search for a more care-full aesthetics, but it does suggest that ‘actively seeking out opportunities to be creative together’ (67) might be a good starting point where the show is not always the thing. I would argue that there is a boldness and important aesthetic quality in work that ‘seeks no external recognition’ (66) because it implies that aesthetic value is found in co-created moments and not only in public display. The execution of a project figured around an aesthetic of care, therefore, relies on building mutual activities of sharing, support, co-working and relational solidarity within a framework of artistry or creative endeavour. Aesthetic value is located in-between people in moments of collaborative creation, conjoined effort and intimate exchange: these are new virtuosities of care that do not rely on the singular display of self-honed skill.
This then returns me to the question of an aesthetics of care. If as I have already argued a care based ethics helps raise questions of justice and ‘other regarding effort’, how might art making be judged from this perspective? The starting point is the notion of relations and the simplistic statement that art making takes place in a series of relational acts, some more explicit and intentional than others. Where an ethics of care focuses upon the values inherent, exhibited or perhaps desired within these human interdependencies, the aesthetics of care seeks to focus upon how the sensory and affective are realised in human relations fostered in art projects. The French art theorist Nicolas Bourriaud is a useful point of departure here in his work on relational aesthetics. Bourriaud defines a relational aesthetic as a ‘set of artistic practices which take as their theoretical and practical point of departure the whole of human relations and their social context, rather than an independent and private space’ (2011, 113) and his book announced from the perspective of the late 1990s French visual art scene how ‘for some years now, there has been an upsurge of convivial, user-friendly artistic projects, festive, collective and participatory, exploring the varied potential in the relationship to the other’ (61). Bourriaud’s work appears to attach an implicit value to this upsurge, with the ‘angelic programme’ I mentioned at the top of the article being instigated in order ‘to patiently re-stitch the relational fabric’ (36) and ‘turn the beholder into the neighbour’ (43). However, ultimately he is more concerned with the formal aspects of this trajectory than the potential that new relational practices have for announcing or creating a fairer world. Inventing new neighbourly relations today, he asserts, is disconnected from programmes that seek to foster ‘happier tomorrows’ (45). As the writing above on the ethics of care would suggest, and the contract of mutual regard urged by Geras, the values that can be materialised in the convivial should in fact be the ground on which happier tomorrows are built. The power of the concept of relational aesthetics is weakened by the fact it does not suggest why relations with others might be endorsed or what type of relations we might aspire to develop. If the socially critical inference in the notion of relational aesthetics is to have greater explicit ethical weight, and to move from a moderately interesting description of a form to a movement with more normative clout, it needs to be refigured. And the argument here is that thinking in terms of an aesthetics of care might provide this reorientation. An ‘aesthetics of care’ is then about a set of values realised in a relational process that emphasise engagements between individuals or groups over time. It is one that might consist of small creative encounters or large-scale exhibitions, but it is always one that notices inter-human relations in both the creation and the display of art projects. It is an aesthetics that is unafraid to lay bare what Shannon Jackson calls the ‘supporting infrastructures of [ … ] living beings’ (2011, 39), but importantly this is an aesthetics that could both present those mutually beneficial structures and foster them. It would not pretend to a distinction between a process and an outcome because both might stimulate affective solidarity between people – perhaps participant to participant or performer to audience. There is a sense that this aesthetics would value intimacy, but it would not be at the expense of what Nato Thompson refers to as ‘explicitly local, long-term, and community-based’ engagement (2012, 31). While care might be exhibited fleetingly, it is more likely that care aesthetics would be realised in more enduring, crafted encounters between people. Seeking to overcome widespread social indifference implies commitment to deep and extended processes. At the beginning of the article it was noted that the ethics of care is a reference to both a set of values and a practice. This is repeated in this proposal for an aesthetics of care, so that it suggests both a demonstration of mutual regard, but simultaneously it instigates a process that is seeking to create or secure it. Amin asserts that there is a process of cultivation in his project for overcoming the disregard experienced in a society of strangers. It is a ‘craft that requires continual attentiveness and care, such that empathy – for objects, projects, nature, the commons – can spread as a public sentiment that also serves to regulate feelings among strangers’ (2012, 7). Attentiveness (a term also found in Tronto (2013, 34)) is both at the heart of the creative process, and the outcome of it. An aesthetics of care is, therefore, a sensory ethical practice, that following Robinson, involves ‘not only learning how to be attentive and patient, how to listen and respond, but also how to rethink our own attitudes about difference and exclusion’ (1999, 164). The difficulty in nurturing an ethics of care through an aesthetic process should of course be acknowledged. Kester in talking about participatory modes of art making explains it as a ‘temporally extensive form of social interaction in which models of expression, enunciation, and reception are continuously modified and reciprocally responsive’ (2011, 112). It is a form of crafted caring where learning to create, respond and be in close dialogue with others is vital for the quality of the experience: but it is a ‘temporally extensive form’ because it needs to be ‘continuously modified’ as it is practiced. In order to outline something of the shape of that form, the remainder of this article will divide care aesthetics into preparation, execution and exhibition, where each of these moments might be minutely connected but they are also likely to be dispersed over a long period. First, preparation would involve an openness and honesty of intention, the selection of artists or participants and questions of the location of a project. Decisions about accessibility (whether in terms of the appropriateness of the space for disabled people, the location in terms of costs of travel or the timing for people with different commitments) are not mundane organisational matters, but crucial ethical propositions. In being taken in reference to the ethics of care, they will imbue the project with an affective sense of the importance of mutual respect and regard. Jackson’s ‘supporting infrastructures’ are not the hidden mechanism of creative endeavour but a valued component of the aesthetics. Preparation is, therefore, paradoxically part of the exhibition within this mode of artistic project: it can demonstrate and model a form of mutual regard. There is a sensory quality in the relationships to which a project that prepares in this way aspires. The notion of execution focuses on the process of collaborative working on artistic projects which forge inter-human relationships. The emerging connections between individuals coalescing in this process have an aesthetics – a shape, feel, sensation and affect. This does not exist within one particular person or object of the work, but appears in-between those involved, so that there is a sensory quality of the process and outcome that cannot be disaggregated from the collective effort. This connects to Richard Sennett’s work on the history and practice of cooperation. His conviction that a practice of working together, and his central example of the Hull House settlement in Chicago, demonstrates that a shared commitment to building caring relations turns ‘people outward in shared, symbolic acts’ and these in turn have a potential place in countering a society that is figured ‘brutally simple: us-againstthem coupled with you-are-on-your-own’ (Sennett 2012, 280). For theatre this form of cooperation might involve a challenge to the quality and texture of a rehearsal or devising experience so that the reciprocity of gradual creation is valued over and above the discipline of the single-minded voyage towards the first night. Those intimate negotiations are the aesthetics of the project, and not merely an unremarkable preparatory period. Borrowing from David Gauntlett’s work on craft, there is a suggestion that within a creative process the power is realised ‘in gentle and quiet ways, with no need for grand celebratory announcements’ (2011, 66). This is not meant as a dismissal of the art of public theatre making as part of the search for a more care-full aesthetics, but it does suggest that ‘actively seeking out opportunities to be creative together’ (67) might be a good starting point where the show is not always the thing. I would argue that there is a boldness and important aesthetic quality in work that ‘seeks no external recognition’ (66) because it implies that aesthetic value is found in co-created moments and not only in public display. The execution of a project figured around an aesthetic of care, therefore, relies on building mutual activities of sharing, support, co-working and relational solidarity within a framework of artistry or creative endeavour. Aesthetic value is located in-between people in moments of collaborative creation, conjoined effort and intimate exchange: these are new virtuosities of care that do not rely on the singular display of self-honed skill.
考点1:relational acts 推荐翻译为“关系实践” 考点2:‘other regarding effort’ 推荐翻译为“关怀他人的努力” 考点3:supporting infrastructures 推荐翻译为“支撑性基础设施” 考点4:user-friendly artistic projects 推荐翻译为“公众易参与的艺术项目”或“亲和型艺术实践” 考点5:Hull House 推荐翻译为“赫尔馆” 考点6:relational fabric 推荐翻译为“关系纽带” 考点7:theater 推荐翻译为“戏剧” 考点8:Nicolas Bourriaud 推荐翻译为“尼古拉斯·伯瑞奥德” 考点9:Geras (David Geras) 推荐翻译为“吉拉斯” 考点10:Tronto (Joan C. Tronto) 推荐翻译为“特朗托” 考点11:Richard Sennett 推荐翻译为“理查德·桑内特” 考点12:David Gauntlett 推荐翻译为“戴维·冈特莱特” 考点13:Nato Thompson 推荐翻译为“内托·汤普森” 考点14:reciprocity推荐翻译为“交互性” 考点15:useful point of departure 推荐译为“一个切实的出发点”,不可译为“一个实用的出发” 考点16:"temporally extensive form"推荐译为“短暂深入的过程” 考点17:"figured"推荐译为“被塑造”
1cfdf
学术论文
人文科学
167
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 第78届戛纳电影节于周六在法国落幕,伊朗异议导演贾法·帕纳希(Jafar Panahi)凭借地下拍摄的复仇题材影片《那只是个意外》(It Was Just an Accident)摘得金棕榈奖,引发全场起立鼓掌。这位因政治立场曾被伊朗当局软禁和入狱的导演,首次在15年内获准亲自到场领奖。 该片讲述五名前政治犯与当年监狱施暴者重逢的故事,由帕纳希的真实囚禁经历启发创作。在发表感言时,帕纳希呼吁全球伊朗人团结起来:“最重要的是我们的国家和自由。”影片反映他在伊文监狱的经历,也象徵着对压迫制度的反抗。 帕纳希曾因出席绿色革命遇难学生葬礼遭伊朗政府禁拍、禁行15年。尽管如此,他持续以秘密方式拍片,作品曾通过USB偷偷送出伊朗。本次夺冠的影片正是在这样的背景下拍摄而成,堪称其生涯的巅峰之一。 评审团特别奖授予中国导演毕赣的《狂野时代》,这是一部融合历史与科幻元素的电影,讲述一个女人被带到了末世后的未来,在那里,她试图修复一个半机器半人的男人,向他隐喻讲述中国历史的部分内容。之后,她必须做出选择,是回到现实世界,还是继续与这个仿生人独处。 本届电影节不仅是一场电影盛宴,也是政治表达的舞台。在美国总统特朗普宣称考虑对海外影片征收100%关税后,不少导演发出批评声音。此外,评审团主席朱丽叶·比诺什(Juliette Binoche)在颁奖典礼上表示:“艺术终将胜利,人性终将胜利。” 电影节闭幕当天,戛纳所在地区遭遇大规模停电,导致交通信号灯和ATM瘫痪,电影节组织方紧急启动备用发电机维持晚间活动顺利进行。据当地检察院称,位于尼斯和戛纳之间的小镇卢贝镇的一座高压电线塔的四根支柱中有三根被锯断。由于损坏,电网运营商不得不关闭该线路。这间接导致了戛纳大规模停电。此外,坦内隆镇的一个变电站也发生火灾。法国当局怀疑这两起案件都是蓄意破坏。调查工作正在进行之中。 **获奖名单——** 主竞赛单元 金棕榈最佳影片:《普通事故》,贾法·帕纳西导演 评委会大奖:《情感价值》,约阿希姆·提尔导演 最佳女演员:纳迪亚·梅利蒂《最小的女儿》 最佳导演:小克莱伯·门多萨《秘密特工》 评审团奖(双黄):《接近终点》,奥利维尔·拉克谢导演&《望向太阳》 ,玛莎·施林斯基导演 最佳编剧:达内兄弟《年轻母亲之家》 最佳男演员:瓦格纳·马拉《秘密特工》 特别奖:《狂野时代》毕赣 短片单元 短片金棕榈:《我很高兴你现在死了》,托菲克·巴霍姆导演 短片特别提及:《阿里》,阿德南·拉杰夫导演 金摄影机(表彰首部导演作品) 金摄影机奖:《总统的蛋糕》,哈桑·哈迪导演 金摄影机奖特别提及:《父影之下》,阿基诺拉·戴维斯导演
第78届戛纳电影节于周六在法国落幕,伊朗异议导演贾法·帕纳希(Jafar Panahi)凭借地下拍摄的复仇题材影片《那只是个意外》(It Was Just an Accident)摘得金棕榈奖,引发全场起立鼓掌。这位因政治立场曾被伊朗当局软禁和入狱的导演,首次在15年内获准亲自到场领奖。 该片讲述五名前政治犯与当年监狱施暴者重逢的故事,由帕纳希的真实囚禁经历启发创作。在发表感言时,帕纳希呼吁全球伊朗人团结起来:“最重要的是我们的国家和自由。”影片反映他在伊文监狱的经历,也象徵着对压迫制度的反抗。 帕纳希曾因出席绿色革命遇难学生葬礼遭伊朗政府禁拍、禁行15年。尽管如此,他持续以秘密方式拍片,作品曾通过USB偷偷送出伊朗。本次夺冠的影片正是在这样的背景下拍摄而成,堪称其生涯的巅峰之一。 评审团特别奖授予中国导演毕赣的《狂野时代》,这是一部融合历史与科幻元素的电影,讲述一个女人被带到了末世后的未来,在那里,她试图修复一个半机器半人的男人,向他隐喻讲述中国历史的部分内容。之后,她必须做出选择,是回到现实世界,还是继续与这个仿生人独处。 本届电影节不仅是一场电影盛宴,也是政治表达的舞台。在美国总统特朗普宣称考虑对海外影片征收100%关税后,不少导演发出批评声音。此外,评审团主席朱丽叶·比诺什(Juliette Binoche)在颁奖典礼上表示:“艺术终将胜利,人性终将胜利。” 电影节闭幕当天,戛纳所在地区遭遇大规模停电,导致交通信号灯和ATM瘫痪,电影节组织方紧急启动备用发电机维持晚间活动顺利进行。据当地检察院称,位于尼斯和戛纳之间的小镇卢贝镇的一座高压电线塔的四根支柱中有三根被锯断。由于损坏,电网运营商不得不关闭该线路。这间接导致了戛纳大规模停电。此外,坦内隆镇的一个变电站也发生火灾。法国当局怀疑这两起案件都是蓄意破坏。调查工作正在进行之中。 **获奖名单——** 主竞赛单元 金棕榈最佳影片:《普通事故》,贾法·帕纳西导演 评委会大奖:《情感价值》,约阿希姆·提尔导演 最佳女演员:纳迪亚·梅利蒂《最小的女儿》 最佳导演:小克莱伯·门多萨《秘密特工》 评审团奖(双黄):《接近终点》,奥利维尔·拉克谢导演&《望向太阳》 ,玛莎·施林斯基导演 最佳编剧:达内兄弟《年轻母亲之家》 最佳男演员:瓦格纳·马拉《秘密特工》 特别奖:《狂野时代》毕赣 短片单元 短片金棕榈:《我很高兴你现在死了》,托菲克·巴霍姆导演 短片特别提及:《阿里》,阿德南·拉杰夫导演 金摄影机(表彰首部导演作品) 金摄影机奖:《总统的蛋糕》,哈桑·哈迪导演 金摄影机奖特别提及:《父影之下》,阿基诺拉·戴维斯导演
考点1:“复仇题材影片”推荐译为 revenge-themed film / revenge drama,不可简单译为 revenge movie,避免语体不匹配。 考点2:“象征着对压迫制度的反抗”推荐译为 a symbol of resistance against the oppressive regime,需注意表达政治含义,不可弱化为 opposition 或 fight back。 考点3:电影名翻译必须正确,《狂野时代》(Resurrection),官方译法 考点4:电影名翻译必须正确,《情感价值》(Sentimental Value),官方译法 考点5:电影名翻译必须正确,《最小的女儿》(The Little Sister),官方译法 考点6:电影名翻译必须正确,《秘密特工》(The Secret Agent),官方译法 考点7:电影名翻译必须正确,《接近终点》(Sirat),官方译法 考点8:电影名翻译必须正确,《望向太阳》(Sound of Falling),官方译法 考点9:电影名翻译必须正确,《年轻母亲之家》(Young Mothers),官方译法 考点10:电影名翻译必须正确,《我很高兴你现在死了》(I'm Glad You're Dead Now),官方译法 考点11:电影名翻译必须正确,《阿里》(Ali),官方译法 考点12:电影名翻译必须正确,《总统的蛋糕》(The President's Cake),官方译法 考点13:电影名翻译必须正确,《父影之下》(My Father's Shadow)),官方译法 考点14:“艺术终将胜利,人性终将胜利” 的原文为 “Art will always win. What is human will always win.” 译文的理想状态是能把这句话还原。
1db4e
新闻资讯
新闻报道
21
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: 8. Travelling & Living: Customer’s personnel shall be visiting at Supplier’s Site for training &Acceptance /reviews etc. and Supplier shall provide assistance in provisioning of invitation letter for obtaining visa. However, all sorts of expenses on account of travelling, boarding/lodging etc shall be borne by Supplier as per clause 5.d. This item will be Optional and Customer will inform Supplier later. If it is arranged by Supplier, Supplier will receive the payment as indicated in the contract at Appendix VI – Part IV (breakup of cost mentioned) without submitting any receipts. 9. Product Deliverables: All the products and deliverables related to the Equipment are listed in Annex 1A/1B, Annex 2B. 10. Program Management: The Program Management will be executed in accordance with the Program Management Plan detailed in the Management Proposal attached in Annex 2B. 11. Training: 11.1. SUPPLIER will provide training to 12x Customer trainees for 22x working days at SUPPLIER’s site and 04x working days training at CUSTOMER site. The purpose of the training is to provide CUSTOMER with necessary information related to the operation of the Equipment, its functionalities and analysis of first level issues as provided in Annex 2B. Several training sessions are foreseen. They shall be performed according to the Delivery Schedule. 11.2. For sake of clarity, SUPPLIER shall bear all costs and expenses of its trainers such as salaries and social security contributions. 11.3. All courses, lectures, practical work exercises and technical demonstrations shall be conducted in English. 11.4. The trainees shall have sufficient knowledge in the field in which they are to be trained. CUSTOMER will provide a full-time interpreter for all sessions to support the trainees who do not have sufficient working knowledge of the English language. 11.5. SUPPLIER shall perform the trainings in accordance with its customary and usual standards. Notwithstanding the here above, in no case shall SUPPLIER be liable in case the trainees have not reached the objectives of the training. 12. Delivery Schedule: SUPPLIER shall make all reasonable efforts to perform the Works according to the Delivery Schedule. The installation, On Site Acceptance Tests and Support Services shall begin at latest two (2) Week after notification by CUSTOMER of availability of the Equipment at the Site. It shall be performed without interruption. CUSTOMER hereby acknowledges that performance by SUPPLIER of the Works depends on the proper and timely performance by CUSTOMER of its obligations under the Contract. Therefore, in case of any delay in the performance by CUSTOMER of any of its obligations under the Contract which affects or causes delay in the performance of the Works by SUPPLIER and/or its subcontractor, SUPPLIER shall be authorized to extend the time period for performance of any part of the Works by the same duration of the delay and/or consequences of that delay on SUPPLIER and /or subcontractor’s performance of Work. CUSTOMER shall not be entitled to claim any liquidated damages to SUPPLIER in such event. 13. Coming into force – Duration The Effective date of the Contract – EDC - (T0) is the date on which all the conditions below are fulfilled: a. Both the parties have signed the Contract. b. Receipt of sample imagery to CUSTOMER from SUPPLIER for quality assessment & successful qualification of imagery and the acceptance of sample imagery on the basis of quality assessment & successful qualification of imagery by Customer shall be a pre-requisite for declaration of EDC. c. The Supplier has obtained all the authorizations, licenses and Export Permits from respective Governments. However, if the Export Permits, Authorizations and Licenses are not required from the respective Government, The SUPPLIER shall submit a declaration certificate duly signed and stamped which is attached at Appendix XIV. d. Each Party shall inform the other Party without undue delay about reception of any necessary approvals and authorizations. If any of the conditions precedent is not satisfied within 6 (six) months from the date of signature of the Contract or such later date as the Parties may agree, any of the Parties may rescind the Contract by written notice to the other without any liability to the other. 14. Acceptance/ Factory Readiness Review: The CUSTOMER shall visit SELLER/SUPPLIER site at the Factory Readiness Review as an observer. The Factory Readiness Review shall be performed in the SUPPLIER’s subcontractor premises. The Factory Readiness Review Certificate (Appendix-I/B) shall be signed by the CUSTOMER and SUPPLIER after successful completion of tests mentioned in Appendix-I/A. Once the Factory Readiness Review Certificate is signed by both parties, it shall constitute the SUPPLIER’s determination that the Equipment has been manufactured in accordance with Annex 2B and can be delivered. SUPPLIER will not ship non-conforming Equipment or an Equipment different from the Equipment as accepted in the Factory Readiness Review. Supplier shall provide Acceptance Test Parameters (Appendix-I/A) at EDC+1 month for customer’s review. 15. Delivery and Transfer of Title and Risks: 15.1 Delivery terms Once the conditions required to deliver the Equipment are met, as provided in Annex 2B, the Equipment shall be delivered according to ICC Incoterms 2020 DAP to Customer ISLAMABAD FINAL INSTALLATION SITE. All transportation, shipping and Insurance shall be conducted under the provisions of such Inco terms. SUPPLIER shall utilize standard packing to ship the Equipment. As soon as the Equipment is delivered for dispatch, fax / e-mail information shall be sent to CUSTOMER by SUPPLIER immediately, giving following details: 1) Name of airline indicating AWB number and date. 2) Name of city from where the consignment is being dispatched. 3) Date of shipment. 4) Expected date of arrival at airport of Destination Country. 5) Name of shipping agents at destination (if any). 6) Total number of packages being dispatched with gross weight and volume of each package. In accordance of the DAP (ICC Incoterms 2020), any penalty levied on CUSTOMER by the Custom authorities etc during clearance of the consignments, due to wrong/incomplete documentation or supplied quantity which is caused by Supplier, will be recovered from Supplier. 15.2 Dispatch of Documents: The soft copies of following documents shall be provided by the Supplier immediately after the customs clearance in China, and two hard copies shall be delivered by courier against the payment remittance 1) Duly signed invoice showing value of the items actually dispatched. 2) Packing list. 3) Duly signed Airway Bill indicating actual freight paid. 4) Original Warranty/Guarantee/Mill Certificate (if applicable). 5) Guarantee certificate that correct goods as per order have been dispatched 6) Duly signed and stamped Insurance Policy/Cover Note covering all risks up to Final Installation Customer Site at Islamabad, Pakistan The Shipment of the complete system(s) shall be done through AIR. Preferably, the consignment should be dispatched through Pakistani Flag Carrier or otherwise with the consent of Customer. 15.3 Transfer of Risk Risk of loss with respect to the Equipment shall be transferred from Supplier to Customer upon Equipment delivery at Customer Islamabad Final Installation Site. As from the transfer of risks, all losses sustained by the Equipment, as well as injury to people or damage to belongings, are at the CUSTOMER’s expense and its sole liability. Upon delivery, CUSTOMER undertakes to ensure that the Equipment is transported and stored in a secured environment and in accordance with usual requirements of transport and storage requirements. 15.4 Title to goods Without prejudice to the provisions of Article “Intellectual Property Rights”, the title to the Equipment is acquired upon complete payment of the Equipment. Notwithstanding the foregoing, it is prohibited to sell, transfer, loan Equipment to any third parties, without prior written consent of Supplier. 16. Shipment and Insurance: Supplier is responsible for the costs of shipping (including Insurance) and transport of the Equipment until Customer Islamabad Final Installation Site according to Incoterm DAP 2020.The charges of transportation (including of insurance) shall be paid at the price as indicated in the contract Appendix-VI and Appendix-VII. Customer is responsible for all Pakistani import taxes, value added tax, customs duties, fees, or duty charges or the cost of incoming handling by Pakistani customs or import agents and any other associated costs for securing release of the shipment from customs. 17. On-Site Acceptance The On Site Acceptance Tests shall be performed on Site after installation of the Equipment, according to the Delivery Schedule. SUPPLIER shall provide detailed procedures of FRR and OSAT to CUSTOMER for approval, two weeks prior to start of these activities. SUPPLIER shall confirm to CUSTOMER the date for commencement of the On Site Acceptance Tests, with a Seven (7) Days prior notice in writing (by mail, fax or email). CUSTOMER shall have an authorized representative present for the entire duration of the On Site Acceptance Tests. The On Site Acceptance Certificate shall be issued by SUPPLIER and signed by CUSTOMER within ten working (10) Days upon demonstration that the Equipment is operating without Anomaly. This timeframe passed, the OSAC is deemed signed and the OSAT is deemed passed. In case the On Site Acceptance Certificate could not be issued because of a Anomaly, SUPPLIER shall, at its own cost, correct the defect according to a time schedule to be agreed upon in writing by the Parties, and SUPPLIER shall perform again the On Site Acceptance Tests. The signature of the On Site Acceptance Certificate shall constitute the authorization for CUSTOMER to use the Equipment, and shall be the starting point of the Equipment warranty. Once the On Site Acceptance Certificate is signed by both parties, payment will be made within 60 calendar days by CUSTOMER to SUPPLIER as per Appendix-VII. 18. Warranty Service: Supplier, at its sole cost and expense shall provide to CUSTOMER the warranty service of two (2) Years on the Equipment (Ground Segment) starting on the date of signing the On Site Acceptance Certificate and also provide duly sign and stamped Supplier’s Standard Warranty as per Appendix-VIII. The warranty service shall be provided by Supplier free of cost to ensure the Equipment/System (Ground Segment) operates without Anomaly, and Supplier shall exercise the skill, care and diligence in the said services that is customary and standard in the industry for such services and equipment. The warranty service defined in Annex 2B constitute the sole and entire liability incurred by the SUPPLIER in event of any Anomaly of the Equipment. For Space Segment, Supplier shall provide the free of cost warranty of imagery services for 3 years i.e. T0+3 years. Furthermore, in space segment Supplier shall also provide the In-Orbit Test Services and 5 years calibration and after sales technical support starting from T0 and lasts till T0+ 5 years.
8. Travelling & Living: Customer’s personnel shall be visiting at Supplier’s Site for training &Acceptance /reviews etc. and Supplier shall provide assistance in provisioning of invitation letter for obtaining visa. However, all sorts of expenses on account of travelling, boarding/lodging etc shall be borne by Supplier as per clause 5.d. This item will be Optional and Customer will inform Supplier later. If it is arranged by Supplier, Supplier will receive the payment as indicated in the contract at Appendix VI – Part IV (breakup of cost mentioned) without submitting any receipts. 9. Product Deliverables: All the products and deliverables related to the Equipment are listed in Annex 1A/1B, Annex 2B. 10. Program Management: The Program Management will be executed in accordance with the Program Management Plan detailed in the Management Proposal attached in Annex 2B. 11. Training: 11.1. SUPPLIER will provide training to 12x Customer trainees for 22x working days at SUPPLIER’s site and 04x working days training at CUSTOMER site. The purpose of the training is to provide CUSTOMER with necessary information related to the operation of the Equipment, its functionalities and analysis of first level issues as provided in Annex 2B. Several training sessions are foreseen. They shall be performed according to the Delivery Schedule. 11.2. For sake of clarity, SUPPLIER shall bear all costs and expenses of its trainers such as salaries and social security contributions. 11.3. All courses, lectures, practical work exercises and technical demonstrations shall be conducted in English. 11.4. The trainees shall have sufficient knowledge in the field in which they are to be trained. CUSTOMER will provide a full-time interpreter for all sessions to support the trainees who do not have sufficient working knowledge of the English language. 11.5. SUPPLIER shall perform the trainings in accordance with its customary and usual standards. Notwithstanding the here above, in no case shall SUPPLIER be liable in case the trainees have not reached the objectives of the training. 12. Delivery Schedule: SUPPLIER shall make all reasonable efforts to perform the Works according to the Delivery Schedule. The installation, On Site Acceptance Tests and Support Services shall begin at latest two (2) Week after notification by CUSTOMER of availability of the Equipment at the Site. It shall be performed without interruption. CUSTOMER hereby acknowledges that performance by SUPPLIER of the Works depends on the proper and timely performance by CUSTOMER of its obligations under the Contract. Therefore, in case of any delay in the performance by CUSTOMER of any of its obligations under the Contract which affects or causes delay in the performance of the Works by SUPPLIER and/or its subcontractor, SUPPLIER shall be authorized to extend the time period for performance of any part of the Works by the same duration of the delay and/or consequences of that delay on SUPPLIER and /or subcontractor’s performance of Work. CUSTOMER shall not be entitled to claim any liquidated damages to SUPPLIER in such event. 13. Coming into force – Duration The Effective date of the Contract – EDC - (T0) is the date on which all the conditions below are fulfilled: a. Both the parties have signed the Contract. b. Receipt of sample imagery to CUSTOMER from SUPPLIER for quality assessment & successful qualification of imagery and the acceptance of sample imagery on the basis of quality assessment & successful qualification of imagery by Customer shall be a pre-requisite for declaration of EDC. c. The Supplier has obtained all the authorizations, licenses and Export Permits from respective Governments. However, if the Export Permits, Authorizations and Licenses are not required from the respective Government, The SUPPLIER shall submit a declaration certificate duly signed and stamped which is attached at Appendix XIV. d. Each Party shall inform the other Party without undue delay about reception of any necessary approvals and authorizations. If any of the conditions precedent is not satisfied within 6 (six) months from the date of signature of the Contract or such later date as the Parties may agree, any of the Parties may rescind the Contract by written notice to the other without any liability to the other. 14. Acceptance/ Factory Readiness Review: The CUSTOMER shall visit SELLER/SUPPLIER site at the Factory Readiness Review as an observer. The Factory Readiness Review shall be performed in the SUPPLIER’s subcontractor premises. The Factory Readiness Review Certificate (Appendix-I/B) shall be signed by the CUSTOMER and SUPPLIER after successful completion of tests mentioned in Appendix-I/A. Once the Factory Readiness Review Certificate is signed by both parties, it shall constitute the SUPPLIER’s determination that the Equipment has been manufactured in accordance with Annex 2B and can be delivered. SUPPLIER will not ship non-conforming Equipment or an Equipment different from the Equipment as accepted in the Factory Readiness Review. Supplier shall provide Acceptance Test Parameters (Appendix-I/A) at EDC+1 month for customer’s review. 15. Delivery and Transfer of Title and Risks: 15.1 Delivery terms Once the conditions required to deliver the Equipment are met, as provided in Annex 2B, the Equipment shall be delivered according to ICC Incoterms 2020 DAP to Customer ISLAMABAD FINAL INSTALLATION SITE. All transportation, shipping and Insurance shall be conducted under the provisions of such Inco terms. SUPPLIER shall utilize standard packing to ship the Equipment. As soon as the Equipment is delivered for dispatch, fax / e-mail information shall be sent to CUSTOMER by SUPPLIER immediately, giving following details: 1) Name of airline indicating AWB number and date. 2) Name of city from where the consignment is being dispatched. 3) Date of shipment. 4) Expected date of arrival at airport of Destination Country. 5) Name of shipping agents at destination (if any). 6) Total number of packages being dispatched with gross weight and volume of each package. In accordance of the DAP (ICC Incoterms 2020), any penalty levied on CUSTOMER by the Custom authorities etc during clearance of the consignments, due to wrong/incomplete documentation or supplied quantity which is caused by Supplier, will be recovered from Supplier. 15.2 Dispatch of Documents: The soft copies of following documents shall be provided by the Supplier immediately after the customs clearance in China, and two hard copies shall be delivered by courier against the payment remittance 1) Duly signed invoice showing value of the items actually dispatched. 2) Packing list. 3) Duly signed Airway Bill indicating actual freight paid. 4) Original Warranty/Guarantee/Mill Certificate (if applicable). 5) Guarantee certificate that correct goods as per order have been dispatched 6) Duly signed and stamped Insurance Policy/Cover Note covering all risks up to Final Installation Customer Site at Islamabad, Pakistan The Shipment of the complete system(s) shall be done through AIR. Preferably, the consignment should be dispatched through Pakistani Flag Carrier or otherwise with the consent of Customer. 15.3 Transfer of Risk Risk of loss with respect to the Equipment shall be transferred from Supplier to Customer upon Equipment delivery at Customer Islamabad Final Installation Site. As from the transfer of risks, all losses sustained by the Equipment, as well as injury to people or damage to belongings, are at the CUSTOMER’s expense and its sole liability. Upon delivery, CUSTOMER undertakes to ensure that the Equipment is transported and stored in a secured environment and in accordance with usual requirements of transport and storage requirements. 15.4 Title to goods Without prejudice to the provisions of Article “Intellectual Property Rights”, the title to the Equipment is acquired upon complete payment of the Equipment. Notwithstanding the foregoing, it is prohibited to sell, transfer, loan Equipment to any third parties, without prior written consent of Supplier. 16. Shipment and Insurance: Supplier is responsible for the costs of shipping (including Insurance) and transport of the Equipment until Customer Islamabad Final Installation Site according to Incoterm DAP 2020.The charges of transportation (including of insurance) shall be paid at the price as indicated in the contract Appendix-VI and Appendix-VII. Customer is responsible for all Pakistani import taxes, value added tax, customs duties, fees, or duty charges or the cost of incoming handling by Pakistani customs or import agents and any other associated costs for securing release of the shipment from customs. 17. On-Site Acceptance The On Site Acceptance Tests shall be performed on Site after installation of the Equipment, according to the Delivery Schedule. SUPPLIER shall provide detailed procedures of FRR and OSAT to CUSTOMER for approval, two weeks prior to start of these activities. SUPPLIER shall confirm to CUSTOMER the date for commencement of the On Site Acceptance Tests, with a Seven (7) Days prior notice in writing (by mail, fax or email). CUSTOMER shall have an authorized representative present for the entire duration of the On Site Acceptance Tests. The On Site Acceptance Certificate shall be issued by SUPPLIER and signed by CUSTOMER within ten working (10) Days upon demonstration that the Equipment is operating without Anomaly. This timeframe passed, the OSAC is deemed signed and the OSAT is deemed passed. In case the On Site Acceptance Certificate could not be issued because of a Anomaly, SUPPLIER shall, at its own cost, correct the defect according to a time schedule to be agreed upon in writing by the Parties, and SUPPLIER shall perform again the On Site Acceptance Tests. The signature of the On Site Acceptance Certificate shall constitute the authorization for CUSTOMER to use the Equipment, and shall be the starting point of the Equipment warranty. Once the On Site Acceptance Certificate is signed by both parties, payment will be made within 60 calendar days by CUSTOMER to SUPPLIER as per Appendix-VII. 18. Warranty Service: Supplier, at its sole cost and expense shall provide to CUSTOMER the warranty service of two (2) Years on the Equipment (Ground Segment) starting on the date of signing the On Site Acceptance Certificate and also provide duly sign and stamped Supplier’s Standard Warranty as per Appendix-VIII. The warranty service shall be provided by Supplier free of cost to ensure the Equipment/System (Ground Segment) operates without Anomaly, and Supplier shall exercise the skill, care and diligence in the said services that is customary and standard in the industry for such services and equipment. The warranty service defined in Annex 2B constitute the sole and entire liability incurred by the SUPPLIER in event of any Anomaly of the Equipment. For Space Segment, Supplier shall provide the free of cost warranty of imagery services for 3 years i.e. T0+3 years. Furthermore, in space segment Supplier shall also provide the In-Orbit Test Services and 5 years calibration and after sales technical support starting from T0 and lasts till T0+ 5 years.
考点1:liquidated damages 必须译为违约金 考点2:claim 建议翻译为请求赔偿/索赔 考点3:a pre-requisite 建议翻译为先决条件/前提条件 考点4:authorizations、 licenses建议翻译为核准文件、执照 考点5:a declaration certificate建议翻译为声明书/确认书 考点6:transfer of Title and Risks 必须译为所有权和风险转移 考点7:DAP建议翻译为目的地交货,不能保留原文不译 考点8:clearance of the consignments 建议翻译为货物清关 考点9:Original Warranty/Guarantee/Mill Certificate建议翻译为原始保修/质保/出厂检验证书 考点10:cover note 必须译为承保条 考点11:Pakistani Flag Carrier必须译为巴基斯坦国家航空公司或巴基斯坦载旗航空公司 考点12:duty charges 建议翻译为对货物征收的税 考点13:Factory Readiness Tests 建议翻译为出厂准备就绪评审 考点14:Factory Readiness Review Certificate 建议翻译为出厂准备就绪证明/确认 考点15:care and diligence 建议翻译为谨慎及勤勉
1e13c
垂类场景
法律
146
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Herman Melville was Hawthorne’s good friend, also an important novelist. Melville's greatest work, Moby Dick (1851) was based on Melville's adventures on the whaling ships. It is the deep "tragedies of human thought" that show his critical understanding of human nature. Today Melville is considered one of America's greatest writers today. Romanism was extremely influenced in a rising America as America had always had a strong spiritual tradition and romanticism was very comfortable with American spiritual heritage and its ideals of democracy and equality. During this period, the American literature was so changeable that has never been before. Nathaniel Hawthorne, Herman Melville, Walt Whitman and Edgar Allan Poe, these four great writers had given depth and strength to American literature at that time. 4. Local Color And Realism: A New Flavor to American Literature After the Civil War, the American society was in a turbulent situation through several economical crises. The writings about critical realism and unveiling the dark side of the society were increased. They were mainly focus on bankrupt in countries, difficult life or struggle of low-position people and so on. Thus Romanism was on the wane with passing days, while Realism rose and became more and more popular. (1) Local Color Fiction and Mark Twain Local Color Fiction first appeared in the early 19th century, and it had further developing after the Civil War. This kind of literature mainly describes the local life. Its keynote was optimistic, and the language was narrative and humorous. For example, the work of Bret Harte (1836-1902) told us the life of American western miners. Mark Twain was the main writer of this period. He wrote for nearly 50 years, and he had actually written many different types of stories. Nevertheless, Twain is remembered most for The Adventures of Tom Sawyer (1876), Life on the Mississippi (1884) and The Adventures of Huckleberry Finn (1884). The characters he created were humorous and full of wittiness. Mark Twain’s work was regarded the witness of America’s pure local life. According to Calkins, “Few American writers have written the same after reading Twain, for he has helped change the entire country with his humor and skillful story telling.” (Carroll C.Calkins, 124). (2) Rise of Realism “Realism is the theory of writing in which familiar aspects of contemporary life and everyday scenes represented in a straightforward or matter of fact manner“ (Wu Dingo, 59). Romantic writers focused on the development of plot, make the story as interesting and attractive as possible, while realism writers emphasized the characterization of characters, focused on objectivity rather than an idealistic view of human nature and human experience. William Dean Howells (1837-1920) insisted that Realism was the truthful treatment of material moral problems of society. And in O.Henry (1862-1910)’s fictions, all his characters were common people and always had an ironical and surprised ending. There was another style of writing developed by Henry James (1843-1916), who was a writer focused on the description of psychology and behavior. He wrote some complex and profound novels such as The Wings of the Dove, The Ambassador, The Golden Bowl, and created psychological realism. After the mid-19th century, the keynote of romantic literature changed form optimistic to doubtful. The sharp conflict of society force more writers’ attention to the unveiling of dark social fact and self-questioning. Thought the writers of this time unveiled the misfortune and sadness or even degenerate things in actual life, they didn’t just took of the dark side of life as their material, most of them were the reformists with the hope of helping to create a better nation. Stephen Crane (1871-1900) showed more and more serious problems in big sites in his Maggie, A Girl of the Streets, and his most famous book about the American Civil War called The Red Badge of Courage. He thought his works liked a mirror reflecting all life, he emphasized the accidental physiological nature of the characters rather than their moral and rational qualities. Stephen Crane had formed a new style called Naturalism, which had influenced many following writers. 5. Modern American Literature The First World War not only damaged the people’s life, but also led to a turbulent situation of the American society. People of this time were named “The Lost Generation ”, and the writers and works had a pessimistic and disconsolate feeling. (1) Modern poetry Ezra Pound, T.S.Eliot and E.E Cummings are three poets who opened the way to modern poetry. Ezra Pound started the “Imagist” movement, and his The Love Song of J. Alfred Prufrock has been called the first masterpiece of modernism. The Waste Land of T.S.Eliot particularly comments on the inhumanity and decadence of large modern cities. (2) Modern novels Many persons regarded that Ernest Hemingway and other important writers of 20th century had adopted the concise style and naturalism of Stephen Crane. Nevertheless, they still created their own styles and had written so many immortal masterpieces. Among the greatest were Ernest Hemingway (l899-1961) and William Faulkner (l897-1962). Ernest Hemingway was once take part in the First Would War, so many of his works deal with war or injury, and nearly all of them examined the nature of courag, e. By suffering from the violent of war, he felt that he was cut off from all his old beliefs and assumptions about life. “He thought the War had broken America’s culture and traditions, and separated it from its toots” (Elisabeth B. Booz: 1982). The works he wrote--The Sun Also Rises, A Farewell to Arms, For Whom the Bell Tolls and The Old Man and the Sea—inferred the state of mind, and they became the representatives of the feeling of this generation. Epilogue American literature has gone though the progress of development over 200 years. It is characterized by the distinct individualism, which is optimistic, free and always creative. The living American literature has been providing potent thinking headsprings for the writers past and nowadays, and it will continue reanimating the talents to bequeath and enrich the tradition of American literature, of which deserved to be proud. Bibliography: 1. Carroll C.Calkins, The story of America, New York: The Readers Digest Association, Inc, 1975 2. Elisabeth B. Booz, A Brief Introduction To Modern American Literature 1919-1980, Shanghai: Shanghai Foreign Language Education Press, 1982 3. 柯恩, Landmarks of American Literature, 北京: 三联书店, 1988 4. Wu Dingbo, An outline of American literature, Shanghai: Shanghai Foreign Language Education Press, 1998 5. Edited by W. P. Trent, J. Erskine & S. P. Sherman, The Cambridge History of English and American Literature, Cambridge, England: University Press, 1997
Herman Melville was Hawthorne’s good friend, also an important novelist. Melville's greatest work, Moby Dick (1851) was based on Melville's adventures on the whaling ships. It is the deep "tragedies of human thought" that show his critical understanding of human nature. Today Melville is considered one of America's greatest writers today. Romanism was extremely influenced in a rising America as America had always had a strong spiritual tradition and romanticism was very comfortable with American spiritual heritage and its ideals of democracy and equality. During this period, the American literature was so changeable that has never been before. Nathaniel Hawthorne, Herman Melville, Walt Whitman and Edgar Allan Poe, these four great writers had given depth and strength to American literature at that time. 4. Local Color And Realism: A New Flavor to American Literature After the Civil War, the American society was in a turbulent situation through several economical crises. The writings about critical realism and unveiling the dark side of the society were increased. They were mainly focus on bankrupt in countries, difficult life or struggle of low-position people and so on. Thus Romanism was on the wane with passing days, while Realism rose and became more and more popular. (1) Local Color Fiction and Mark Twain Local Color Fiction first appeared in the early 19th century, and it had further developing after the Civil War. This kind of literature mainly describes the local life. Its keynote was optimistic, and the language was narrative and humorous. For example, the work of Bret Harte (1836-1902) told us the life of American western miners. Mark Twain was the main writer of this period. He wrote for nearly 50 years, and he had actually written many different types of stories. Nevertheless, Twain is remembered most for The Adventures of Tom Sawyer (1876), Life on the Mississippi (1884) and The Adventures of Huckleberry Finn (1884). The characters he created were humorous and full of wittiness. Mark Twain’s work was regarded the witness of America’s pure local life. According to Calkins, “Few American writers have written the same after reading Twain, for he has helped change the entire country with his humor and skillful story telling.” (Carroll C.Calkins, 124). (2) Rise of Realism “Realism is the theory of writing in which familiar aspects of contemporary life and everyday scenes represented in a straightforward or matter of fact manner“ (Wu Dingo, 59). Romantic writers focused on the development of plot, make the story as interesting and attractive as possible, while realism writers emphasized the characterization of characters, focused on objectivity rather than an idealistic view of human nature and human experience. William Dean Howells (1837-1920) insisted that Realism was the truthful treatment of material moral problems of society. And in O.Henry (1862-1910)’s fictions, all his characters were common people and always had an ironical and surprised ending. There was another style of writing developed by Henry James (1843-1916), who was a writer focused on the description of psychology and behavior. He wrote some complex and profound novels such as The Wings of the Dove, The Ambassador, The Golden Bowl, and created psychological realism. After the mid-19th century, the keynote of romantic literature changed form optimistic to doubtful. The sharp conflict of society force more writers’ attention to the unveiling of dark social fact and self-questioning. Thought the writers of this time unveiled the misfortune and sadness or even degenerate things in actual life, they didn’t just took of the dark side of life as their material, most of them were the reformists with the hope of helping to create a better nation. Stephen Crane (1871-1900) showed more and more serious problems in big sites in his Maggie, A Girl of the Streets, and his most famous book about the American Civil War called The Red Badge of Courage. He thought his works liked a mirror reflecting all life, he emphasized the accidental physiological nature of the characters rather than their moral and rational qualities. Stephen Crane had formed a new style called Naturalism, which had influenced many following writers. 5. Modern American Literature The First World War not only damaged the people’s life, but also led to a turbulent situation of the American society. People of this time were named “The Lost Generation ”, and the writers and works had a pessimistic and disconsolate feeling. (1) Modern poetry Ezra Pound, T.S.Eliot and E.E Cummings are three poets who opened the way to modern poetry. Ezra Pound started the “Imagist” movement, and his The Love Song of J. Alfred Prufrock has been called the first masterpiece of modernism. The Waste Land of T.S.Eliot particularly comments on the inhumanity and decadence of large modern cities. (2) Modern novels Many persons regarded that Ernest Hemingway and other important writers of 20th century had adopted the concise style and naturalism of Stephen Crane. Nevertheless, they still created their own styles and had written so many immortal masterpieces. Among the greatest were Ernest Hemingway (l899-1961) and William Faulkner (l897-1962). Ernest Hemingway was once take part in the First Would War, so many of his works deal with war or injury, and nearly all of them examined the nature of courag, e. By suffering from the violent of war, he felt that he was cut off from all his old beliefs and assumptions about life. “He thought the War had broken America’s culture and traditions, and separated it from its toots” (Elisabeth B. Booz: 1982). The works he wrote--The Sun Also Rises, A Farewell to Arms, For Whom the Bell Tolls and The Old Man and the Sea—inferred the state of mind, and they became the representatives of the feeling of this generation. Epilogue American literature has gone though the progress of development over 200 years. It is characterized by the distinct individualism, which is optimistic, free and always creative. The living American literature has been providing potent thinking headsprings for the writers past and nowadays, and it will continue reanimating the talents to bequeath and enrich the tradition of American literature, of which deserved to be proud. Bibliography: 1. Carroll C.Calkins, The story of America, New York: The Readers Digest Association, Inc, 1975 2. Elisabeth B. Booz, A Brief Introduction To Modern American Literature 1919-1980, Shanghai: Shanghai Foreign Language Education Press, 1982 3. 柯恩, Landmarks of American Literature, 北京: 三联书店, 1988 4. Wu Dingbo, An outline of American literature, Shanghai: Shanghai Foreign Language Education Press, 1998 5. Edited by W. P. Trent, J. Erskine & S. P. Sherman, The Cambridge History of English and American Literature, Cambridge, England: University Press, 1997
考点1:"was very comfortable with"应译为“与…十分契合”。 考点2:"on the wane'推荐译为“日渐式微”。 考点3:"Local Color Fiction"应译为“乡土小说”。 考点4:"keynote"应译为“基调”。
24d78
学术论文
人文科学
44
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 人工智能医疗应用中的算法责任认定与数据合规边界 一、算法决策的可解释性争议 人工智能在医疗领域的深度应用(如影像辅助诊断、临床决策支持系统)引发了 “算法黑箱” 与医疗责任的核心矛盾。根据《欧盟人工智能法案》,医疗 AI 被归类为 “高风险系统”,要求其决策过程具备 “可追溯性”与 “人机协同验证”。但实践中存在三重冲突: 1. 技术特性与法律要求的张力 深度学习模型(如卷积神经网络 CNN)的 “非线性运算” 特性导致决策逻辑难以拆解,而《通用数据保护条例》第 22 条赋予用户 “解释权”,要求企业说明自动化决策的依据。例如,2023 年某 AI 辅助诊断系统误诊早期肺癌,患者家属要求医院解释 “算法为何将磨玻璃结节判定为良性”,但医院以 “模型参数属于商业秘密” 为由拒绝,引发合规争议。 2. 责任划分的模糊地带 当 AI 决策与医师判断冲突时,责任归属存在歧义:若医师完全依赖 AI 结论导致误诊,属 “医疗过失”;若医师未采纳 AI 的正确预警而造成损害,是否构成 “不作为侵权”?2024 年某三甲医院案例中,AI 提示 “患者术后大出血风险极高”,主治医师未予处理,最终患者死亡,法院判定 “医师对 AI 建议负有审慎审查义务”,但未明确 AI 研发方的连带责任。 3. 跨境应用的标准冲突 美国 FDA 对医疗 AI 采用 “产品分级监管”(如将影像诊断 AI 归类为 二级医疗器械),侧重 “性能验证”;中国 NMPA 则要求 “全生命周期管理”,包括算法迭代的备案更新。某跨国企业的 AI 心电分析系统在美通过审批后,因未在中国完成算法版本变更备案,被认定为 “非法行医辅助工具”,凸显标准差异的执行困境。 二、医疗数据跨境传输的合规障碍 医疗数据的 “敏感性”使其跨境传输受严格限制,核心争议集中于: 1. 数据本地化与出境例外的博弈 中国《数据安全法》要求 “重要数据本地化存储”,而 “国际多中心临床试验” 常需向境外传输病例数据。某药企为加速肿瘤新药研发,向欧盟合作方传输 500 例患者基因测序数据,因未通过 “安全评估” 被处以 2000 万元罚款。欧盟则允许 “基于充分性认定” 的自由传输,但尚未将中国纳入 “白名单”,企业需依赖 “标准合同条款”完成合规,流程繁琐。 2. 匿名化与去标识化的边界混淆 GDPR 规定 “匿名化数据不受监管”,而 “去标识化数据仍属个人数据”。实践中,医疗机构常误将 “去除姓名、身份证号” 的病历视为 “匿名化”,实则通过 “出生日期 + 疾病史” 可重新识别个体。2023 年某健康 APP 因向境外传输 “去标识化糖尿病患者数据” 被欧盟认定为 “数据泄露”,面临全球营收 4% 的罚款。 3. 紧急医疗场景的合规例外缺失 跨国救援中,实时传输患者电子健康档案可能违反数据出境规定。2024 年某国际航班突发心梗病例,机组向美国地面医院传输患者既往病史数据,事后被质疑 “未事先获得患者授权”,但现行法律未明确 “紧急情况豁免” 条款,导致医疗行为陷入 “救人与合规” 的两难。 三、前沿问题:生成式 AI 在医疗中的伦理与法律挑战 ChatGPT 等生成式 AI 在病历总结、诊疗建议中的应用,引发新的合规风险: 1. 幻觉输出的医疗风险 生成式 AI 可能产生 “看似合理却错误” 的内容(如虚构药物禁忌症)。某基层医院使用 AI 撰写出院小结时,误将 “阿司匹林过敏” 写入无过敏史患者的病历,导致后续用药错误。此类 “算法幻觉”是否构成 “产品缺陷”,现行法律尚未明确责任主体。 2. 知识产权的归属争议 AI 基于海量医学文献生成的诊疗方案,是否侵犯原作者的 “著作权”?某 AI 公司的肿瘤治疗建议系统因大段借鉴《NCCN 指南》内容,被起诉 “不正当竞争”,法院认为 “实质性相似但未构成侵权”,但未界定 “合理使用” 的边界。 3. 患者知情同意的形式缺陷 使用 AI 辅助诊疗需获得患者 “专项同意”,而非笼统的 “诊疗同意书”。某互联网医院在用户注册协议中默认勾选 “允许 AI 分析病情”,被认定为 “未充分告知”,违反《个人信息保护法》第 14 条,面临整改。 四、争议解决路径的创新探索 针对上述困境,实践中出现三类解决方案: 1. 算法备案与第三方审计 中国推行 “医疗 AI 算法备案制”,要求企业提交 “训练数据来源、决策逻辑流程图”;欧盟则要求高风险 AI 系统接受 “独立合规评估”。某 AI 眼底筛查系统通过第三方验证其 “误诊率 < 0.5%”,成为首个同时通过中欧合规认证的医疗 AI 产品。 2. 数据信托与跨境白名单 新加坡建立 “医疗数据信托”,由第三方机构统一管理跨境传输,平衡 “隐私保护” 与 “科研需求”;中国粤港澳大湾区试点 “数据跨境流动白名单”,允许指定医院向港澳传输脱敏病例数据,无需单独审批。 3. 替代性纠纷解决机制 国际医疗 AI 纠纷逐渐采用 “专家仲裁”,如国际医学人工智能协会设立仲裁院,由医师、数据合规师、算法工程师组成合议庭,2024 年成功调解某跨国 AI 诊断系统的跨境责任纠纷,耗时较诉讼缩短 60%。
人工智能医疗应用中的算法责任认定与数据合规边界 一、算法决策的可解释性争议 人工智能在医疗领域的深度应用(如影像辅助诊断、临床决策支持系统)引发了 “算法黑箱” 与医疗责任的核心矛盾。根据《欧盟人工智能法案》,医疗 AI 被归类为 “高风险系统”,要求其决策过程具备 “可追溯性”与 “人机协同验证”。但实践中存在三重冲突: 1. 技术特性与法律要求的张力 深度学习模型(如卷积神经网络 CNN)的 “非线性运算” 特性导致决策逻辑难以拆解,而《通用数据保护条例》第 22 条赋予用户 “解释权”,要求企业说明自动化决策的依据。例如,2023 年某 AI 辅助诊断系统误诊早期肺癌,患者家属要求医院解释 “算法为何将磨玻璃结节判定为良性”,但医院以 “模型参数属于商业秘密” 为由拒绝,引发合规争议。 2. 责任划分的模糊地带 当 AI 决策与医师判断冲突时,责任归属存在歧义:若医师完全依赖 AI 结论导致误诊,属 “医疗过失”;若医师未采纳 AI 的正确预警而造成损害,是否构成 “不作为侵权”?2024 年某三甲医院案例中,AI 提示 “患者术后大出血风险极高”,主治医师未予处理,最终患者死亡,法院判定 “医师对 AI 建议负有审慎审查义务”,但未明确 AI 研发方的连带责任。 3. 跨境应用的标准冲突 美国 FDA 对医疗 AI 采用 “产品分级监管”(如将影像诊断 AI 归类为 二级医疗器械),侧重 “性能验证”;中国 NMPA 则要求 “全生命周期管理”,包括算法迭代的备案更新。某跨国企业的 AI 心电分析系统在美通过审批后,因未在中国完成算法版本变更备案,被认定为 “非法行医辅助工具”,凸显标准差异的执行困境。 二、医疗数据跨境传输的合规障碍 医疗数据的 “敏感性”使其跨境传输受严格限制,核心争议集中于: 1. 数据本地化与出境例外的博弈 中国《数据安全法》要求 “重要数据本地化存储”,而 “国际多中心临床试验” 常需向境外传输病例数据。某药企为加速肿瘤新药研发,向欧盟合作方传输 500 例患者基因测序数据,因未通过 “安全评估” 被处以 2000 万元罚款。欧盟则允许 “基于充分性认定” 的自由传输,但尚未将中国纳入 “白名单”,企业需依赖 “标准合同条款”完成合规,流程繁琐。 2. 匿名化与去标识化的边界混淆 GDPR 规定 “匿名化数据不受监管”,而 “去标识化数据仍属个人数据”。实践中,医疗机构常误将 “去除姓名、身份证号” 的病历视为 “匿名化”,实则通过 “出生日期 + 疾病史” 可重新识别个体。2023 年某健康 APP 因向境外传输 “去标识化糖尿病患者数据” 被欧盟认定为 “数据泄露”,面临全球营收 4% 的罚款。 3. 紧急医疗场景的合规例外缺失 跨国救援中,实时传输患者电子健康档案可能违反数据出境规定。2024 年某国际航班突发心梗病例,机组向美国地面医院传输患者既往病史数据,事后被质疑 “未事先获得患者授权”,但现行法律未明确 “紧急情况豁免” 条款,导致医疗行为陷入 “救人与合规” 的两难。 三、前沿问题:生成式 AI 在医疗中的伦理与法律挑战 ChatGPT 等生成式 AI 在病历总结、诊疗建议中的应用,引发新的合规风险: 1. 幻觉输出的医疗风险 生成式 AI 可能产生 “看似合理却错误” 的内容(如虚构药物禁忌症)。某基层医院使用 AI 撰写出院小结时,误将 “阿司匹林过敏” 写入无过敏史患者的病历,导致后续用药错误。此类 “算法幻觉”是否构成 “产品缺陷”,现行法律尚未明确责任主体。 2. 知识产权的归属争议 AI 基于海量医学文献生成的诊疗方案,是否侵犯原作者的 “著作权”?某 AI 公司的肿瘤治疗建议系统因大段借鉴《NCCN 指南》内容,被起诉 “不正当竞争”,法院认为 “实质性相似但未构成侵权”,但未界定 “合理使用” 的边界。 3. 患者知情同意的形式缺陷 使用 AI 辅助诊疗需获得患者 “专项同意”,而非笼统的 “诊疗同意书”。某互联网医院在用户注册协议中默认勾选 “允许 AI 分析病情”,被认定为 “未充分告知”,违反《个人信息保护法》第 14 条,面临整改。 四、争议解决路径的创新探索 针对上述困境,实践中出现三类解决方案: 1. 算法备案与第三方审计 中国推行 “医疗 AI 算法备案制”,要求企业提交 “训练数据来源、决策逻辑流程图”;欧盟则要求高风险 AI 系统接受 “独立合规评估”。某 AI 眼底筛查系统通过第三方验证其 “误诊率 < 0.5%”,成为首个同时通过中欧合规认证的医疗 AI 产品。 2. 数据信托与跨境白名单 新加坡建立 “医疗数据信托”,由第三方机构统一管理跨境传输,平衡 “隐私保护” 与 “科研需求”;中国粤港澳大湾区试点 “数据跨境流动白名单”,允许指定医院向港澳传输脱敏病例数据,无需单独审批。 3. 替代性纠纷解决机制 国际医疗 AI 纠纷逐渐采用 “专家仲裁”,如国际医学人工智能协会设立仲裁院,由医师、数据合规师、算法工程师组成合议庭,2024 年成功调解某跨国 AI 诊断系统的跨境责任纠纷,耗时较诉讼缩短 60%。
考点 1:“不作为侵权” 必须译为 “tort of omission”,专业术语,固定译法 考点 2:“国家药品监督管理局(NMPA)” 必须译为 “National Medical Products Administration (NMPA)”,国家机关官方译名,不可译为其他 考点 3:“独立合规评估” 推荐译为 “independent conformity assessment”,不可译为 “independent compliance evaluation”或“compliance assessment” 考点4:“基层医院”推荐译为“primary care hospital”或“community hospital”,不可译为“grassroots hospital” 考点5: “诊疗建议”推荐译为“diagnostic and treatment advice”或更概括的“clinical advice” 考点6:"决策逻辑难以拆解”中的“拆解”应为“deconstruct”或“interpret”,不可译为“disassemble” 考点7:“数据本地化与出境例外的博弈”中的“博弈”一词应译为“tension”、“interplay”或“balancing act”,不可译为“game”
25168
学术论文
社会科学
182
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 近些年,尽管白象食品在线上渠道取得亮眼成绩和令友商们羡慕的增速。但白象主营方便面产品口味欠佳,创新产品又频频翻车,猎奇口味如香菜面和折耳根拌面遭到差评,高端线“鲜面传”复购率低,致使消费者对白象产品的信任度下降。 北京消费者王女士称,经常刷到白象香菜方便面的推广视频,层层香菜包裹着面条的画面对于一个香菜爱好者而言充满了好奇和购买欲望,6·18大促前夕,她花了39.9元拍下10包,吃着却觉得口味一般,面条偏咸,王女士表示不想再买第二次。 与此同时,在预制菜、外卖以及越来越多样化的方便速食新品冲击下,原本基数庞大的方便面消费群体逐渐转向螺狮粉、热干面、预制菜等市场,方便面整体市场几近天花板。据世界方便面协会数据,中国方便面消费量已从2020年巅峰期的472.3亿份降至2023年的431.2亿份。 另有《方便面市场消费量的影响因素分析》提供的数据,外卖市场规模每增加1%,方便面消费量便减0.0533%;同时,自热火锅、螺蛳粉等新型方便速食和预制菜的崛起,都在分流方便面消费。 在此背景下,对于业务相对单一、饮料市场尚未形成气候,仍聚焦面食的白象食品而言,更是雪上加霜。为了拓展新的增长点,不少方便面厂家纷纷瞄准“超大泡面”,试图通过打造多元化和满足消费者对性价比要求进行破局。比如,2024年下半年今麦郎推出“超大泡面”,面饼和配料总重430克,包含四块面饼,主打四人份。 2024年白象也推出“超大泡面”新品,定位“多人食分享装”,提供香菜面和火鸡面等多种口味。今年2月,白象再度推出新品——汤好喝超大桶。 但就在近日,多位网友反映称,白象多半袋面、多半桶面系列产品包装上显示“多半袋”为注册商标,实际面饼仅增加35克(如普通版70克增至110-120克),增幅不足50%,并质疑企业“在宣传上玩文字游戏”。白象食品官方客服人员也向媒体表示,“多半袋”确实是商标,产品具体克重以包装上显示的为准,本身产品是没有问题的。 事件曝光后,“玩文字游戏”话题登上热搜,消费者集中质疑企业诚信。白象次日紧急致歉,承认“初衷是为区分常规份量”,承诺调整包装以避免误解。但在业内人士看来,此类营销会破坏良性市场竞争,导致消费者对行业信任度下降。 也有法律专家指出,多数消费者将“多半袋面”理解为“分量比普通包装多一半”,此类表述可能违反《广告法》中“不得含有虚假或引人误解内容”的规定,利用商标权规避虚假宣传责任的做法存在法律风险。 对于白象而言,渠道管理失衡与价格体系失控后的战争已然打响,目前市场竞争和消费信任也正面临前所未有的挑战,未来该何去何从,尚待市场检验。
近些年,尽管白象食品在线上渠道取得亮眼成绩和令友商们羡慕的增速。但白象主营方便面产品口味欠佳,创新产品又频频翻车,猎奇口味如香菜面和折耳根拌面遭到差评,高端线“鲜面传”复购率低,致使消费者对白象产品的信任度下降。 北京消费者王女士称,经常刷到白象香菜方便面的推广视频,层层香菜包裹着面条的画面对于一个香菜爱好者而言充满了好奇和购买欲望,6·18大促前夕,她花了39.9元拍下10包,吃着却觉得口味一般,面条偏咸,王女士表示不想再买第二次。 与此同时,在预制菜、外卖以及越来越多样化的方便速食新品冲击下,原本基数庞大的方便面消费群体逐渐转向螺狮粉、热干面、预制菜等市场,方便面整体市场几近天花板。据世界方便面协会数据,中国方便面消费量已从2020年巅峰期的472.3亿份降至2023年的431.2亿份。 另有《方便面市场消费量的影响因素分析》提供的数据,外卖市场规模每增加1%,方便面消费量便减0.0533%;同时,自热火锅、螺蛳粉等新型方便速食和预制菜的崛起,都在分流方便面消费。 在此背景下,对于业务相对单一、饮料市场尚未形成气候,仍聚焦面食的白象食品而言,更是雪上加霜。为了拓展新的增长点,不少方便面厂家纷纷瞄准“超大泡面”,试图通过打造多元化和满足消费者对性价比要求进行破局。比如,2024年下半年今麦郎推出“超大泡面”,面饼和配料总重430克,包含四块面饼,主打四人份。 2024年白象也推出“超大泡面”新品,定位“多人食分享装”,提供香菜面和火鸡面等多种口味。今年2月,白象再度推出新品——汤好喝超大桶。 但就在近日,多位网友反映称,白象多半袋面、多半桶面系列产品包装上显示“多半袋”为注册商标,实际面饼仅增加35克(如普通版70克增至110-120克),增幅不足50%,并质疑企业“在宣传上玩文字游戏”。白象食品官方客服人员也向媒体表示,“多半袋”确实是商标,产品具体克重以包装上显示的为准,本身产品是没有问题的。 事件曝光后,“玩文字游戏”话题登上热搜,消费者集中质疑企业诚信。白象次日紧急致歉,承认“初衷是为区分常规份量”,承诺调整包装以避免误解。但在业内人士看来,此类营销会破坏良性市场竞争,导致消费者对行业信任度下降。 也有法律专家指出,多数消费者将“多半袋面”理解为“分量比普通包装多一半”,此类表述可能违反《广告法》中“不得含有虚假或引人误解内容”的规定,利用商标权规避虚假宣传责任的做法存在法律风险。 对于白象而言,渠道管理失衡与价格体系失控后的战争已然打响,目前市场竞争和消费信任也正面临前所未有的挑战,未来该何去何从,尚待市场检验。
考点1:亮眼成绩推荐译为“stellar performance”,而不应译为“bright-eye results”。 考点2:频频翻车推荐译为“repeatedly flopped”。 考点3:猎奇口味推荐译为“exotic flavors”。 考点4:拍下不可直译为直译为“patted down”或“took a picture of”。 考点5:尚未形成气候是一个非常形象的中文成语,比喻某事物还未发展壮大,没有形成规模和影响力,推荐译为“is not yet well-established”,不可直译为“has not yet formed a climate”。 考点6:克重是产品规格术语,应译为“net weight”。 考点7:登上热搜是网络流行语,推荐译为“started trending on social media”。 考点8:尚待市场检验是商业分析报告中常见的结尾套话,推荐译为“remains to be seen”。
25697
新闻资讯
新闻报道
40
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Plastic products derived from petroleum refining have become indispensable in daily life due to their lightweight nature, low production cost, and excellent durability [1]. According to relevant statistics, global plastic consumption has surged from 335 million tons in 2016 to 400 million tons in 2021, showing a continuous upward trend [2]. During the COVID-19 pandemic, the demand for single-use personal protective equipment (PPE) such as masks and gloves, along with packaging materials, surged dramatically. It is estimated that over 450 billion plastic-based PPE items were used globally in 2020 alone [3]. Moreover, the widespread adoption of fast-paced lifestyles has driven the rapid growth in the food delivery and e-commerce sectors, significantly increasing the consumption of plastic packaging materials. However, despite the convenience plastics bring to modern life, the growing dependence on them has led to increasingly severe challenges in waste management. Owing to the large-scale production and short service life of most plastic products, post-use plastics continue to accumulate in the environment, leading to mounting pollution and increasingly complex waste management issues. Currently, the scientific community is actively addressing the challenges of plastic recycling and resource recovery by exploring a range of efficient and viable chemical transformation pathways, including hydrogenolysis, alcoholysis, glycolysis, and pyrolysis [4,5,6]. Under ideal conditions, chemical conversion pathways should enable the efficient recovery of high-purity monomers from plastic waste, with the recovered products exhibiting quality and reactivity comparable to those of virgin feedstocks, allowing for direct reutilization in the synthesis of new polymers. This approach not only significantly improves resource utilization efficiency but also paves the way toward genuine “closed-loop recycling” [7,8]. However, plastics—particularly thermosets and highly cross-linked polymers—exhibit exceptional chemical stability due to their dense three-dimensional network structures and robust carbon–carbon backbones, which severely hinder their degradation under mild conditions. As a result, current chemical recycling technologies typically require harsh operating conditions—such as elevated temperatures (>180 °C), high pressures (20–40 atm), and concentrated alkali (1–10 M) or strong acids (e.g., concentrated H2SO4 or HNO3) as reaction media. While these conditions facilitate polymer-chain scission and improve reaction rates, they also pose significant challenges, including high energy consumption, severe equipment corrosion, secondary environmental pollution, and limited economic feasibility [9,10,11]. Therefore, the development of mild and green plastic conversion technologies operable under ambient or near-ambient conditions is a critical breakthrough toward achieving sustainable plastic recycling. Photo-driven photocatalytic technology is regarded as a highly promising strategy for sustainable plastic waste management. In general, photocatalytic plastic conversion systems can be broadly categorized into the following two types: photodegradation and photosynthetic transformation (Figure 1) [12]. The ability to selectively generate value-added products is one of the key criteria distinguishing photodegradation from selective photo-reforming processes. As shown in Figure 1a, conventional photodegradation typically aims at mineralization, relying on highly oxidative radicals (e.g., •OH) to indiscriminately attack polymer backbones and intermediates. This results in complex reaction pathways and broad product distributions, with final products dominated by CO2, water, and trace organic acids, offering limited value [12,13]. In contrast, photosynthetic conversion emphasizes precise regulation of reaction pathways to efficiently convert hydrocarbon content in plastics into chemically valuable target molecules (e.g., H2, C1–C2+ fuels, and aromatic compounds) (Figure 1b) [14]. Therefore, the transition to photosynthetic systems hinges on the modulation of radical type, concentration, and spatial reactivity to prevent uncontrolled mineralization by overly oxidative species. In this review, we first provide a brief overview of the current status of plastic waste treatment with respect to the basic reaction mechanisms involved in the photocatalytic process for plastics, with special emphasis on the redox pathways and the roles of key reactants in the photocatalytic process. We then systematically summarize representative recent advances in photocatalytic plastic waste conversion, categorized by plastic type (e.g., polyolefins and polyesters) (Figure 2), and critically analyze the performance and applicability of different catalytic systems. Finally, we discuss the key challenges in achieving efficient and selective transformation of plastic waste, including catalyst stability, compatibility with real-world plastics, product selectivity control, and system-level economic feasibility. We also outline future directions in catalyst design, mechanistic investigation, reaction engineering scale-up, and industrial deployment. We hope this review offers theoretical insights and technical references that contribute to the green transition of plastic pollution management and resource recovery. The management of plastic waste has attracted widespread global attention due to its significant impact on the sustainability of human life. A variety of techniques are employed in managing plastic waste, including incineration, landfill disposal, microbial degradation, and mechanical shredding with subsequent recycling, as well as transformation methods driven by electricity, heat, or light. These approaches each offer distinct advantages and limitations in the effort to reduce environmental impact and recover value from discarded plastic materials (Figure 3) [15,16,17,18,19,20,21]. Currently, plastic waste is primarily managed through landfilling and incineration, while a substantial fraction is still discharged directly into natural environments [22]. Among various plastic products, packaging materials are particularly problematic, as they are prone to entering rivers, lakes, and oceans, posing significant threats to aquatic ecosystems [23]. With the continuous growth in global plastic production and the diminishing availability of landfill space, the economic burden of plastic waste management has increased markedly. More critically, landfilling not only leads to a considerable waste of petroleum-derived resources but also gives rise to severe environmental concerns, including the leaching of chemical additives and long-term land occupation [24]. Similarly, although plastic incineration allows partial energy recovery, it is far less efficient compared to recycling and entails high environmental costs [25]. The life-cycle carbon footprint of incinerated polymers is substantial, with an estimated 5 to 10 tons of CO2 emitted per ton of plastic processed [17]. Furthermore, incineration can produce toxic byproducts [26], rendering it a suboptimal strategy from both ecological and energy standpoints. Mechanical recycling typically follows a downcycling approach, wherein waste plastics are physically processed through shredding, washing, and remelting [27]. However, the resulting recycled materials are often utilized in the production of low-performance products, such as flower pots or non-food-contact containers [28]. This degradation in material quality leads to a reduction in overall polymer value, liTo address the inherent limitations of conventional recycling approaches, catalytic upcycling has emerged as a highly attractive alternative. This strategy enables the efficient conversion of polymeric waste into high-value chemicals, fuels, and other functional materials with significant economic potential, offering a promising complement to existing recycling methods [29]. Electrocatalysis has demonstrated promising conversion efficiency in the treatment of polymeric waste, offering advantages such as mild operating conditions and product selectivity. However, its reliance on external electrical energy input and the current limitations in electrode material selection, reactor design, and long-term operational stability pose significant technical challenges that must be addressed to enable practical application [30]. Thermocatalytic conversion is highly efficient and offers excellent reaction stability, enabling the rapid cleavage of C–C bonds to produce value-added chemicals. However, the requirement for elevated temperatures and substantial energy input imposes constraints on its scalability and practical implementation [31]. Enzymatic biotechnology relies on highly specific plastic-degrading enzymes, which have a limited substrate scope and face challenges in processing complex or mixed plastic waste. Moreover, these enzymes often exhibit poor stability under harsh reaction conditions. Fermentation, which converts plastic-derived intermediates into valuable chemicals through microbial metabolism, typically requires tedious pretreatment steps to depolymerize plastics into fermentable substrates. This complicates the process workflow and significantly increases downstream separation and purification costs [32]. Photocatalytic conversion has emerged as a promising strategy for plastic waste treatment due to its mild reaction conditions, high energy efficiency, and potential for selective transformation. As shown in Figure 4, operating under ambient temperature and pressure and utilizing solar or ultraviolet light as the energy source, photocatalysis presents a more economically viable and environmentally friendly alternative compared with conventional thermal and chemical degradation methods [33]. Unlike pyrolysis or incineration, which require high energy input and often generate toxic byproducts, photocatalytic processes can proceed under mild conditions with minimal environmental risk. Furthermore, many photocatalytic systems leverage solar energy, enhancing their long-term sustainability and cost-effectiveness. Certain photocatalysts have demonstrated the ability to selectively cleave polymer chains or oxidize specific moieties, enabling the conversion of plastic waste into valuable monomers or fine chemicals. Compared with enzymatic biotechnology, which relies on highly specific enzymes and suffers from limited substrate scope and low stability under harsh conditions, photocatalytic systems exhibit broader applicability. They can effectively process both homochain and heterochain polymers by modulating reactive oxygen species (•OH, •O2−) and optimizing catalyst active sites. Similarly, microbial fermentation—despite its potential for converting plastic-derived intermediates into chemicals—often requires tedious and costly pretreatment to break down polymers into fermentable substrates, complicating process workflows and increasing separation burdens. In contrast, photocatalysis offers an integrated platform for both degradation and upcycling, enabling direct, selective conversion of plastic waste without extensive preprocessing. Its ability to control radical activity also helps suppress over-oxidation, thereby favoring the formation of high-value products. Taken together, the synergistic advantages of photocatalytic technology, including mild operational conditions, solar energy utilization, broad substrate compatibility, and direct value-added conversion, position it as a highly efficient and sustainable core strategy for addressing plastic waste. In comparison, other emerging alternatives remain constrained by energy intensity, substrate specificity, or process complexity.miting its potential for reuse in high-end applications.
Plastic products derived from petroleum refining have become indispensable in daily life due to their lightweight nature, low production cost, and excellent durability [1]. According to relevant statistics, global plastic consumption has surged from 335 million tons in 2016 to 400 million tons in 2021, showing a continuous upward trend [2]. During the COVID-19 pandemic, the demand for single-use personal protective equipment (PPE) such as masks and gloves, along with packaging materials, surged dramatically. It is estimated that over 450 billion plastic-based PPE items were used globally in 2020 alone [3]. Moreover, the widespread adoption of fast-paced lifestyles has driven the rapid growth in the food delivery and e-commerce sectors, significantly increasing the consumption of plastic packaging materials. However, despite the convenience plastics bring to modern life, the growing dependence on them has led to increasingly severe challenges in waste management. Owing to the large-scale production and short service life of most plastic products, post-use plastics continue to accumulate in the environment, leading to mounting pollution and increasingly complex waste management issues. Currently, the scientific community is actively addressing the challenges of plastic recycling and resource recovery by exploring a range of efficient and viable chemical transformation pathways, including hydrogenolysis, alcoholysis, glycolysis, and pyrolysis [4,5,6]. Under ideal conditions, chemical conversion pathways should enable the efficient recovery of high-purity monomers from plastic waste, with the recovered products exhibiting quality and reactivity comparable to those of virgin feedstocks, allowing for direct reutilization in the synthesis of new polymers. This approach not only significantly improves resource utilization efficiency but also paves the way toward genuine “closed-loop recycling” [7,8]. However, plastics—particularly thermosets and highly cross-linked polymers—exhibit exceptional chemical stability due to their dense three-dimensional network structures and robust carbon–carbon backbones, which severely hinder their degradation under mild conditions. As a result, current chemical recycling technologies typically require harsh operating conditions—such as elevated temperatures (>180 °C), high pressures (20–40 atm), and concentrated alkali (1–10 M) or strong acids (e.g., concentrated H2SO4 or HNO3) as reaction media. While these conditions facilitate polymer-chain scission and improve reaction rates, they also pose significant challenges, including high energy consumption, severe equipment corrosion, secondary environmental pollution, and limited economic feasibility [9,10,11]. Therefore, the development of mild and green plastic conversion technologies operable under ambient or near-ambient conditions is a critical breakthrough toward achieving sustainable plastic recycling. Photo-driven photocatalytic technology is regarded as a highly promising strategy for sustainable plastic waste management. In general, photocatalytic plastic conversion systems can be broadly categorized into the following two types: photodegradation and photosynthetic transformation (Figure 1) [12]. The ability to selectively generate value-added products is one of the key criteria distinguishing photodegradation from selective photo-reforming processes. As shown in Figure 1a, conventional photodegradation typically aims at mineralization, relying on highly oxidative radicals (e.g., •OH) to indiscriminately attack polymer backbones and intermediates. This results in complex reaction pathways and broad product distributions, with final products dominated by CO2, water, and trace organic acids, offering limited value [12,13]. In contrast, photosynthetic conversion emphasizes precise regulation of reaction pathways to efficiently convert hydrocarbon content in plastics into chemically valuable target molecules (e.g., H2, C1–C2+ fuels, and aromatic compounds) (Figure 1b) [14]. Therefore, the transition to photosynthetic systems hinges on the modulation of radical type, concentration, and spatial reactivity to prevent uncontrolled mineralization by overly oxidative species. In this review, we first provide a brief overview of the current status of plastic waste treatment with respect to the basic reaction mechanisms involved in the photocatalytic process for plastics, with special emphasis on the redox pathways and the roles of key reactants in the photocatalytic process. We then systematically summarize representative recent advances in photocatalytic plastic waste conversion, categorized by plastic type (e.g., polyolefins and polyesters) (Figure 2), and critically analyze the performance and applicability of different catalytic systems. Finally, we discuss the key challenges in achieving efficient and selective transformation of plastic waste, including catalyst stability, compatibility with real-world plastics, product selectivity control, and system-level economic feasibility. We also outline future directions in catalyst design, mechanistic investigation, reaction engineering scale-up, and industrial deployment. We hope this review offers theoretical insights and technical references that contribute to the green transition of plastic pollution management and resource recovery. The management of plastic waste has attracted widespread global attention due to its significant impact on the sustainability of human life. A variety of techniques are employed in managing plastic waste, including incineration, landfill disposal, microbial degradation, and mechanical shredding with subsequent recycling, as well as transformation methods driven by electricity, heat, or light. These approaches each offer distinct advantages and limitations in the effort to reduce environmental impact and recover value from discarded plastic materials (Figure 3) [15,16,17,18,19,20,21]. Currently, plastic waste is primarily managed through landfilling and incineration, while a substantial fraction is still discharged directly into natural environments [22]. Among various plastic products, packaging materials are particularly problematic, as they are prone to entering rivers, lakes, and oceans, posing significant threats to aquatic ecosystems [23]. With the continuous growth in global plastic production and the diminishing availability of landfill space, the economic burden of plastic waste management has increased markedly. More critically, landfilling not only leads to a considerable waste of petroleum-derived resources but also gives rise to severe environmental concerns, including the leaching of chemical additives and long-term land occupation [24]. Similarly, although plastic incineration allows partial energy recovery, it is far less efficient compared to recycling and entails high environmental costs [25]. The life-cycle carbon footprint of incinerated polymers is substantial, with an estimated 5 to 10 tons of CO2 emitted per ton of plastic processed [17]. Furthermore, incineration can produce toxic byproducts [26], rendering it a suboptimal strategy from both ecological and energy standpoints. Mechanical recycling typically follows a downcycling approach, wherein waste plastics are physically processed through shredding, washing, and remelting [27]. However, the resulting recycled materials are often utilized in the production of low-performance products, such as flower pots or non-food-contact containers [28]. This degradation in material quality leads to a reduction in overall polymer value, liTo address the inherent limitations of conventional recycling approaches, catalytic upcycling has emerged as a highly attractive alternative. This strategy enables the efficient conversion of polymeric waste into high-value chemicals, fuels, and other functional materials with significant economic potential, offering a promising complement to existing recycling methods [29]. Electrocatalysis has demonstrated promising conversion efficiency in the treatment of polymeric waste, offering advantages such as mild operating conditions and product selectivity. However, its reliance on external electrical energy input and the current limitations in electrode material selection, reactor design, and long-term operational stability pose significant technical challenges that must be addressed to enable practical application [30]. Thermocatalytic conversion is highly efficient and offers excellent reaction stability, enabling the rapid cleavage of C–C bonds to produce value-added chemicals. However, the requirement for elevated temperatures and substantial energy input imposes constraints on its scalability and practical implementation [31]. Enzymatic biotechnology relies on highly specific plastic-degrading enzymes, which have a limited substrate scope and face challenges in processing complex or mixed plastic waste. Moreover, these enzymes often exhibit poor stability under harsh reaction conditions. Fermentation, which converts plastic-derived intermediates into valuable chemicals through microbial metabolism, typically requires tedious pretreatment steps to depolymerize plastics into fermentable substrates. This complicates the process workflow and significantly increases downstream separation and purification costs [32]. Photocatalytic conversion has emerged as a promising strategy for plastic waste treatment due to its mild reaction conditions, high energy efficiency, and potential for selective transformation. As shown in Figure 4, operating under ambient temperature and pressure and utilizing solar or ultraviolet light as the energy source, photocatalysis presents a more economically viable and environmentally friendly alternative compared with conventional thermal and chemical degradation methods [33]. Unlike pyrolysis or incineration, which require high energy input and often generate toxic byproducts, photocatalytic processes can proceed under mild conditions with minimal environmental risk. Furthermore, many photocatalytic systems leverage solar energy, enhancing their long-term sustainability and cost-effectiveness. Certain photocatalysts have demonstrated the ability to selectively cleave polymer chains or oxidize specific moieties, enabling the conversion of plastic waste into valuable monomers or fine chemicals. Compared with enzymatic biotechnology, which relies on highly specific enzymes and suffers from limited substrate scope and low stability under harsh conditions, photocatalytic systems exhibit broader applicability. They can effectively process both homochain and heterochain polymers by modulating reactive oxygen species (•OH, •O2−) and optimizing catalyst active sites. Similarly, microbial fermentation—despite its potential for converting plastic-derived intermediates into chemicals—often requires tedious and costly pretreatment to break down polymers into fermentable substrates, complicating process workflows and increasing separation burdens. In contrast, photocatalysis offers an integrated platform for both degradation and upcycling, enabling direct, selective conversion of plastic waste without extensive preprocessing. Its ability to control radical activity also helps suppress over-oxidation, thereby favoring the formation of high-value products. Taken together, the synergistic advantages of photocatalytic technology, including mild operational conditions, solar energy utilization, broad substrate compatibility, and direct value-added conversion, position it as a highly efficient and sustainable core strategy for addressing plastic waste. In comparison, other emerging alternatives remain constrained by energy intensity, substrate specificity, or process complexity.miting its potential for reuse in high-end applications.
考点1:【single-use personal protective equipment (PPE)】 应译为 【一次性个人防护用品】,其中的“equipment”不可译为“设备” 考点2:【chemical transformation pathways 】应译为 【化学转化路径】 考点3:【glycolysis】 应译为 【乙二醇解】 考点4:【polyolefins and polyesters】 应译为 【聚烯烃与聚酯类】 考点5:【life-cycle carbon footprint】 应译为 【全生命周期碳足迹】 考点6:【reactive oxygen species】 应译为 【活性氧物种】 考点7:【homochain and heterochain polymers】 应译为 【同链/均链与异链/杂链聚合物】 考点8:【green transition of plastic pollution management】 应译为 【塑料污染治理的绿色转型】
25e2e
学术论文
自然科学
87
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Quantitative Transcriptional Analysis of Single-Cell RNA-Sequencing Methods SUMMARY Single-cell RNA sequencing (scRNA-seq) offers new possibilities to address biological and medical questions. However, systematic comparisons of the per- formance of diverse scRNA-seq protocols are lacking. We generated data from 583 mouse embryonic stem cells to evaluate six prominent scRNA-seq methods: CEL-seq2, Drop-seq, MARS-seq, SCRB- seq, Smart-seq, and Smart-seq2. While Smart-seq2 detected the most genes per cell and across cells, CEL-seq2, Drop-seq, MARS-seq, and SCRB-seq quantified mRNA levels with less amplification noise due to the use of unique molecular identifiers (UMIs). Power simulations at different sequencing depths showed that Drop-seq is more cost-efficient for transcriptome quantification of large numbers of cells, while MARS-seq, SCRB-seq, and Smart-seq2 are more efficient when analyzing fewer cells. Our quantitative comparison offers the basis for an informed choice among six prominent scRNA-seq methods, and it provides a framework for benchmarking further improvements of scRNA-seq protocols. INTRODUCTION Genome-wide quantification of mRNA transcripts is highly informative for characterizing cellular states and molecular circuitries (ENCODE Project Consortium, 2012). Ideally, such data are collected with high spatial resolution, and single-cell RNA sequencing (scRNA-seq) now allows for transcriptome-wide analyses of individual cells, revealing exciting biological and med- ical insights (Kolodziejczyk et al., 2015a; Wagner et al., 2016). scRNA-seq requires the isolation and lysis of single cells, the conversion of their RNA into cDNA, and the amplification of cDNA to generate high-throughput sequencing libraries. As the amount of starting material is so small, this process results in substantial technical variation (Kolodziejczyk et al., 2015a; Wagner et al., 2016). One type of technical variable is the sensitivity of a scRNA- seq method (i.e., the probability to capture and convert a particular mRNA transcript present in a single cell into a cDNA molecule present in the library). Another variable of interest is the accuracy (i.e., how well the read quantification corresponds to the actual concentration of mRNAs), and a third type is the precision with which this amplification occurs (i.e., the technical variation of the quantification). The combination of sensitivity, precision, and number of cells analyzed determines the power to detect relative differences in expression levels. Finally, the monetary cost to reach a desired level of power is of high practical relevance. To make a well- informed choice among available scRNA-seq methods, it is important to quantify these parameters comparably. Some strengths and weaknesses of different methods are already known. For example, it has previously been shown that scRNA-seq conducted in the small volumes available in the automated microfluidic platform from Fluidigm (C1 platform) outperforms CEL-seq2, Smart-seq, or other commercially available kits in microliter volumes (Hashimshony et al., 2016; Wu et al., 2014). Furthermore, the Smart-seq protocol has been optimized for sensitivity, more even full-length coverage, accuracy, and cost (Picelli et al., 2013), and this improved Smart-seq2 protocol (Picelli et al., 2014b) has also become widely used (Gokce et al., 2016; Reinius et al., 2016; Tirosh et al., 2016). Other protocols have sacrificed full-length coverage in order to sequence part of the primer used for cDNA generation. This enables early barcoding of libraries (i.e., the incorporation of cell-specific barcodes), allowing for multiplexing the cDNA amplification and thereby increasing the throughput of scRNA- seq library generation by one to three orders of magnitude (Hashimshony et al., 2012; Jaitin et al., 2014; Klein et al., 2015; Macosko et al., 2015; Soumillon et al., 2014). Additionally, this approach allows the incorporation of unique molecular identifiers (UMIs), random nucleotide sequences that tag individual mRNA molecules and, hence, allow for the distinction between original molecules and amplification duplicates that derive from the cDNA or library amplification (Kivioja et al., 2011). Utilization of UMI information improves quantification of mRNA molecules (Grun et al., 2014; Islam et al., 2014), and it has been implemented in several scRNA-seq protocols, such as STRT (Islam et al., 2014), CEL-seq (Gru€n et al., 2014; Hashimshony et al., 2016), CEL-seq2 (Hashimshony et al., 2016), Drop-seq (Ma- cosko et al., 2015), inDrop (Klein et al., 2015), MARS-seq (Jaitin et al., 2014), and SCRB-seq (Soumillon et al., 2014). However, a thorough and systematic comparison of relevant parameters across scRNA-seq methods is still lacking. To address this issue, we generated 583 scRNA-seq libraries from mouse embryonic stem cells (mESCs), using six different methods in two replicates, and we compared their sensitivity, accuracy, precision, power, and efficiency (Figure 1). DISCUSSION Here we have provided an in-depth comparison of six prominent scRNA-seq protocols. To this end, we generated data for all six compared methods from the same cells, cultured under the same condition in the same laboratory. While there would be many more datasets and methods for a comparison of the sensitivity and accuracy of the ERCCs (Svensson et al., 2016), our approach provides a more controlled and comprehensive comparison across thousands of endogenous genes. This is important, as can be seen by the different sensitivity estimates that we obtained for Drop-seq, MARS-seq, and SCRB-seq using the ERCCs. In our comparison, we clearly find that Smart-seq2 is the most sensitive method, closely followed by SCRB-seq, Smart-seq/C1, and CEL-seq2/C1, while Drop-seq and MARS- seq detect nearly 50% fewer genes per cell (Figures 3B and 3C). In addition, Smart-seq2 shows the most even read coverage across transcripts (Figure S3D), making it the most appropriate method for the detection of alternative splice forms and for analyses of allele-specific expression using SNPs (Deng et al., 2014; Reinius et al., 2016). Hence, Smart-seq2 is certainly the most suitable method when an annotation of single-cell transcriptomes is the focus. Furthermore, we find that Smart-seq2 is also the most accurate method (i.e., it has the highest correlation of known ERCC spike-in concentrations and read counts per million), which is probably related to its higher sensitivity. Hence, differences in expression values across transcripts within the same cell predict differences in the actual concentrations of these transcripts well. All methods do this rather well, at least for higher expression levels, and we think that the small differences among methods will rarely be a decisive factor. Importantly, the accuracy of estimating transcript concentrations across cells (relevant, e.g., for comparing the total RNA content of cells) depends on different factors and cannot be compared well among the tested methods as it would require known concentration differences of transcripts across cells. However, it is likely that methods that can use UMIs and ERCCs (CEL-seq2/ C1, MARS-seq, and SCRB-seq) would have a strong advantage in this respect. How well relative expression levels of the same genes can be compared across cells depends on two factors. First, how often (i.e., in how many cells and from how many molecules) it is measured. Second, with how much technical variation (i.e., with how much noise, e.g., from amplification) it is measured. For the first factor (dropout probability), we find Smart-seq2 to be the best method (Figure 5A), as expected from its high gene detection sensitivity. For the second factor (extra Poisson variability), we find the four UMI methods to perform better (Figure 5B), as expected from their ability to eliminate variation introduced by amplification. To assess the combined effect of these two factors, we performed simulations for differential gene expression scenarios (Figure 6). This allowed us to translate the sensitivity and precision parameters into the practically relevant power to detect differentially expressed genes. Of note, our power estimates include the variation that is caused by the two different replicates per method that constitutes an important part of the variation. Our simulations show that, at a sequencing depth of one million reads, SCRB-seq has the highest power, probably due to a good balance of high sensitivity and low amplification noise. Furthermore, amplification noise and power strongly depend on the use of UMIs, especially for the PCR- based methods (Figures 5B and 6B; Figure S7). Notably, this is due to the large amount of amplification needed for scRNA- seq libraries, as the effect of UMIs on power for bulk RNA-seq libraries is negligible (Parekh et al., 2016). Perhaps practically most important, our power simulations also allow us to compare the efficiency of the methods by calculating the costs to generate the data for a given level of power. Using minimal cost calculations, we find that Drop-seq is the most cost-effective method, closely followed by SCRB-seq, MARS-seq, and Smart-seq2. However, Drop-seq costs are likely to be more underestimated, due to lower flexibility in generating a specified number of libraries and the higher fraction of reads that come from bad cells. Hence, all four UMI methods are in practice probably similarly cost-effective. In contrast, for Smart-seq2 to be similarly cost-effective it is absolutely necessary to use in-house-produced transposase or to drastically reduce volumes of commercial transposase kits (Lamble et al., 2013; Mora-Castilla et al., 2016). Given comparable efficiencies of Drop-seq, MARS-seq, SCRB-seq, and Smart-seq2, additional factors will play a role when choosing a suitable method for a particular question. Due to its low library costs, Drop-seq is probably prefer- able when analyzing large numbers of cells at low coverage (e.g., to find rare cell types). On the other hand, Drop-seq in its current setup requires a relatively large amount of cells (>6,500 for 1 min of flow). Hence, if few and/or unstable cells are isolated by FACS, the SCRB-seq, MARS-seq, or Smart- seq2 protocols are probably preferable. Additional advantages of these methods over Drop-seq include that technical variation can be estimated from ERCCs for each cell, which can be helpful to estimate biological variation (Kim et al., 2015; Vallejos et al., 2016), and that the exact same setup can be used to generate bulk RNA-seq libraries. While SCRB-seq is slightly more cost-effective than MARS-seq and has the advantage that one does not need to produce the transposase in-house, Smart-seq2 is preferable when transcriptome annotation, identification of sequence variants, or the quantification of different splice forms is of interest. Furthermore, the presence of batch effects shows that experiments need to be designed in a way that does not confound batches with bio- logical factors (Hicks et al., 2015). Practically, plate-based methods might currently accommodate complex experimental designs with various biological factors more easily than microfluidic chips.
Quantitative Transcriptional Analysis of Single-Cell RNA-Sequencing Methods SUMMARY Single-cell RNA sequencing (scRNA-seq) offers new possibilities to address biological and medical questions. However, systematic comparisons of the per- formance of diverse scRNA-seq protocols are lacking. We generated data from 583 mouse embryonic stem cells to evaluate six prominent scRNA-seq methods: CEL-seq2, Drop-seq, MARS-seq, SCRB- seq, Smart-seq, and Smart-seq2. While Smart-seq2 detected the most genes per cell and across cells, CEL-seq2, Drop-seq, MARS-seq, and SCRB-seq quantified mRNA levels with less amplification noise due to the use of unique molecular identifiers (UMIs). Power simulations at different sequencing depths showed that Drop-seq is more cost-efficient for transcriptome quantification of large numbers of cells, while MARS-seq, SCRB-seq, and Smart-seq2 are more efficient when analyzing fewer cells. Our quantitative comparison offers the basis for an informed choice among six prominent scRNA-seq methods, and it provides a framework for benchmarking further improvements of scRNA-seq protocols. INTRODUCTION Genome-wide quantification of mRNA transcripts is highly informative for characterizing cellular states and molecular circuitries (ENCODE Project Consortium, 2012). Ideally, such data are collected with high spatial resolution, and single-cell RNA sequencing (scRNA-seq) now allows for transcriptome-wide analyses of individual cells, revealing exciting biological and med- ical insights (Kolodziejczyk et al., 2015a; Wagner et al., 2016). scRNA-seq requires the isolation and lysis of single cells, the conversion of their RNA into cDNA, and the amplification of cDNA to generate high-throughput sequencing libraries. As the amount of starting material is so small, this process results in substantial technical variation (Kolodziejczyk et al., 2015a; Wagner et al., 2016). One type of technical variable is the sensitivity of a scRNA- seq method (i.e., the probability to capture and convert a particular mRNA transcript present in a single cell into a cDNA molecule present in the library). Another variable of interest is the accuracy (i.e., how well the read quantification corresponds to the actual concentration of mRNAs), and a third type is the precision with which this amplification occurs (i.e., the technical variation of the quantification). The combination of sensitivity, precision, and number of cells analyzed determines the power to detect relative differences in expression levels. Finally, the monetary cost to reach a desired level of power is of high practical relevance. To make a well- informed choice among available scRNA-seq methods, it is important to quantify these parameters comparably. Some strengths and weaknesses of different methods are already known. For example, it has previously been shown that scRNA-seq conducted in the small volumes available in the automated microfluidic platform from Fluidigm (C1 platform) outperforms CEL-seq2, Smart-seq, or other commercially available kits in microliter volumes (Hashimshony et al., 2016; Wu et al., 2014). Furthermore, the Smart-seq protocol has been optimized for sensitivity, more even full-length coverage, accuracy, and cost (Picelli et al., 2013), and this improved Smart-seq2 protocol (Picelli et al., 2014b) has also become widely used (Gokce et al., 2016; Reinius et al., 2016; Tirosh et al., 2016). Other protocols have sacrificed full-length coverage in order to sequence part of the primer used for cDNA generation. This enables early barcoding of libraries (i.e., the incorporation of cell-specific barcodes), allowing for multiplexing the cDNA amplification and thereby increasing the throughput of scRNA- seq library generation by one to three orders of magnitude (Hashimshony et al., 2012; Jaitin et al., 2014; Klein et al., 2015; Macosko et al., 2015; Soumillon et al., 2014). Additionally, this approach allows the incorporation of unique molecular identifiers (UMIs), random nucleotide sequences that tag individual mRNA molecules and, hence, allow for the distinction between original molecules and amplification duplicates that derive from the cDNA or library amplification (Kivioja et al., 2011). Utilization of UMI information improves quantification of mRNA molecules (Grun et al., 2014; Islam et al., 2014), and it has been implemented in several scRNA-seq protocols, such as STRT (Islam et al., 2014), CEL-seq (Gru€n et al., 2014; Hashimshony et al., 2016), CEL-seq2 (Hashimshony et al., 2016), Drop-seq (Ma- cosko et al., 2015), inDrop (Klein et al., 2015), MARS-seq (Jaitin et al., 2014), and SCRB-seq (Soumillon et al., 2014). However, a thorough and systematic comparison of relevant parameters across scRNA-seq methods is still lacking. To address this issue, we generated 583 scRNA-seq libraries from mouse embryonic stem cells (mESCs), using six different methods in two replicates, and we compared their sensitivity, accuracy, precision, power, and efficiency (Figure 1). DISCUSSION Here we have provided an in-depth comparison of six prominent scRNA-seq protocols. To this end, we generated data for all six compared methods from the same cells, cultured under the same condition in the same laboratory. While there would be many more datasets and methods for a comparison of the sensitivity and accuracy of the ERCCs (Svensson et al., 2016), our approach provides a more controlled and comprehensive comparison across thousands of endogenous genes. This is important, as can be seen by the different sensitivity estimates that we obtained for Drop-seq, MARS-seq, and SCRB-seq using the ERCCs. In our comparison, we clearly find that Smart-seq2 is the most sensitive method, closely followed by SCRB-seq, Smart-seq/C1, and CEL-seq2/C1, while Drop-seq and MARS- seq detect nearly 50% fewer genes per cell (Figures 3B and 3C). In addition, Smart-seq2 shows the most even read coverage across transcripts (Figure S3D), making it the most appropriate method for the detection of alternative splice forms and for analyses of allele-specific expression using SNPs (Deng et al., 2014; Reinius et al., 2016). Hence, Smart-seq2 is certainly the most suitable method when an annotation of single-cell transcriptomes is the focus. Furthermore, we find that Smart-seq2 is also the most accurate method (i.e., it has the highest correlation of known ERCC spike-in concentrations and read counts per million), which is probably related to its higher sensitivity. Hence, differences in expression values across transcripts within the same cell predict differences in the actual concentrations of these transcripts well. All methods do this rather well, at least for higher expression levels, and we think that the small differences among methods will rarely be a decisive factor. Importantly, the accuracy of estimating transcript concentrations across cells (relevant, e.g., for comparing the total RNA content of cells) depends on different factors and cannot be compared well among the tested methods as it would require known concentration differences of transcripts across cells. However, it is likely that methods that can use UMIs and ERCCs (CEL-seq2/ C1, MARS-seq, and SCRB-seq) would have a strong advantage in this respect. How well relative expression levels of the same genes can be compared across cells depends on two factors. First, how often (i.e., in how many cells and from how many molecules) it is measured. Second, with how much technical variation (i.e., with how much noise, e.g., from amplification) it is measured. For the first factor (dropout probability), we find Smart-seq2 to be the best method (Figure 5A), as expected from its high gene detection sensitivity. For the second factor (extra Poisson variability), we find the four UMI methods to perform better (Figure 5B), as expected from their ability to eliminate variation introduced by amplification. To assess the combined effect of these two factors, we performed simulations for differential gene expression scenarios (Figure 6). This allowed us to translate the sensitivity and precision parameters into the practically relevant power to detect differentially expressed genes. Of note, our power estimates include the variation that is caused by the two different replicates per method that constitutes an important part of the variation. Our simulations show that, at a sequencing depth of one million reads, SCRB-seq has the highest power, probably due to a good balance of high sensitivity and low amplification noise. Furthermore, amplification noise and power strongly depend on the use of UMIs, especially for the PCR- based methods (Figures 5B and 6B; Figure S7). Notably, this is due to the large amount of amplification needed for scRNA- seq libraries, as the effect of UMIs on power for bulk RNA-seq libraries is negligible (Parekh et al., 2016). Perhaps practically most important, our power simulations also allow us to compare the efficiency of the methods by calculating the costs to generate the data for a given level of power. Using minimal cost calculations, we find that Drop-seq is the most cost-effective method, closely followed by SCRB-seq, MARS-seq, and Smart-seq2. However, Drop-seq costs are likely to be more underestimated, due to lower flexibility in generating a specified number of libraries and the higher fraction of reads that come from bad cells. Hence, all four UMI methods are in practice probably similarly cost-effective. In contrast, for Smart-seq2 to be similarly cost-effective it is absolutely necessary to use in-house-produced transposase or to drastically reduce volumes of commercial transposase kits (Lamble et al., 2013; Mora-Castilla et al., 2016). Given comparable efficiencies of Drop-seq, MARS-seq, SCRB-seq, and Smart-seq2, additional factors will play a role when choosing a suitable method for a particular question. Due to its low library costs, Drop-seq is probably prefer- able when analyzing large numbers of cells at low coverage (e.g., to find rare cell types). On the other hand, Drop-seq in its current setup requires a relatively large amount of cells (>6,500 for 1 min of flow). Hence, if few and/or unstable cells are isolated by FACS, the SCRB-seq, MARS-seq, or Smart- seq2 protocols are probably preferable. Additional advantages of these methods over Drop-seq include that technical variation can be estimated from ERCCs for each cell, which can be helpful to estimate biological variation (Kim et al., 2015; Vallejos et al., 2016), and that the exact same setup can be used to generate bulk RNA-seq libraries. While SCRB-seq is slightly more cost-effective than MARS-seq and has the advantage that one does not need to produce the transposase in-house, Smart-seq2 is preferable when transcriptome annotation, identification of sequence variants, or the quantification of different splice forms is of interest. Furthermore, the presence of batch effects shows that experiments need to be designed in a way that does not confound batches with bio- logical factors (Hicks et al., 2015). Practically, plate-based methods might currently accommodate complex experimental designs with various biological factors more easily than microfluidic chips.
考点1: “Dropout Probability”必须译为“漏检概率”(非“缺失概率”) 考点2:“unique molecular identifiers”必须译为“独特分子标识符”首现可保留英文 UMI 并加注) 考点3: “quantification”应译为“量化” 考点4: “full-length coverage”应译为“全长覆盖度”(保持“度”的一致性) 考点5: “spike-in“必须译为”掺入对照“(“加标”可备注为行业俗称,但不作首选) 考点6:“Poisson variability”应译为“泊松变异性” 考点7:“>6,500 for 1 min of flow”必须译为“流速>6,500/分钟” 考点8:“plate-based methods”推荐译为“基于微孔板的方法” 考点9: “multiplexing”必须译为“多重化/多路复用”(不应只保留英文) 考点10: “amplification duplicates”必须译为“扩增重复(序列副本)/重复扩增子”(非“扩增产物”) 考点11: “sequence part of the primer”推荐译为“对引物部分进行测序”(非“捕获/捕集”) 考点12: “in how many cells / from how many molecules”推荐译为“在多少细胞中/由多少分子(覆盖范围/数量维度)”(非“被测量的频率”) 考点13: “monetary cost”推荐译为“成本/费用”(非“成本控制”)
263bd
学术论文
自然科学
107
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 2025 年 6 月 16 日,三只松鼠发布公告,宣布终止对湖南爱零食科技有限公司的收购计划。回溯至 2024 年 10 月,三只松鼠全资子公司安徽一件事创业投资有限公司拟斥资不超过 2 亿元,收购爱零食的控制权或相关业务及资产。彼时,双方签署的《投资意向协议》有效期为 90 天,并在今年 1 月将排他期延长至正式签约日,然而最终因部分核心条款未能达成一致,交易宣告流产。 野心勃勃的收购计划:三只松鼠的线下突围战 三只松鼠曾对爱零食寄予厚望,这场收购背后是其迫切的战略需求。从渠道结构来看,长期以来,三只松鼠线上依赖程度颇高。以 2024 年数据为例,其线上销售额占营业收入的比重为 69.73%,达 74.07 亿元 ,尽管同比增长 49.6%,但线下渠道的相对薄弱仍是制约其进一步发展的关键因素。 反观爱零食,作为量贩零食区域龙头,在湖南等多个区域已布局超 1800 家门店。在中国零食市场,2023 年线下销售占据了 82.6% 的市场份额 ,这一庞大的线下市场份额对三只松鼠而言极具吸引力。收购爱零食,无疑能让三只松鼠快速切入线下量贩零食市场,拓展线下版图,优化自身渠道结构,进而提升市场覆盖面和销售规模。 与此同时,量贩零食赛道正上演疯狂扩张。2024 年,零食很忙与赵一鸣零食合并为“鸣鸣很忙”,门店总数迅速突破 10000 家;万辰集团整合好想来、来优品等四大品牌,2025 年签约门店更是突破 1.5 万家。 爱零食在 2024 年 9 月门店数已超 1800 家,并提出“3 年 3000 家店” 的扩张目标,位居行业第五。在这样的竞争环境下,三只松鼠试图通过收购爱零食,在这场竞速赛中抢占一席之地,并借助双方在供应链、品牌上的互补优势,实现资源整合与产品矩阵升级。 财报数据:增长放缓下的战略权衡 三只松鼠 2024 年财报成绩亮眼:营收 106.22 亿元,同比增长 49.30%;归母净利润 4.08 亿元,同比增长 85.51%。线下分销收入也因并购爱折扣增长 48.62% 。但进入 2025 年第一季度,增速骤然放缓,营收仅微增 2.13%,归母净利润下滑 22.46%。 在此背景下,三只松鼠可能重新审视收购爱零食的必要性与可行性。或许是考虑到收购后整合的复杂性与成本,包括供应链整合、门店运营模式融合、人员管理等多方面潜在成本;以及自身业务调整与资源分配的需要。在面对与爱零食核心条款分歧时,选择终止收购。 命运转折:收购终止后的两家企业何去何从? 对于三只松鼠而言,终止收购虽短期内延缓了其在量贩零食赛道通过并购快速扩张的步伐,但也避免了潜在的整合风险与资金投入压力。公司可将资源集中于自有分销网络建设、社区零食店发展以及多品牌孵化等战略布局上。长期来看,如果三只松鼠能有效推进自身线下策略,持续提升供应链优势,进一步扩大线下分销规模,仍有望在零食市场竞争中巩固地位。 对爱零食来说,失去三只松鼠的资金注入与资源协同,门店扩张计划可能受阻。原本计划借助三只松鼠的 2 亿元投资加速门店扩张,目标在 2025 年达到 5000 家店,如今这一计划面临变数。在行业头部品牌竞争加剧的情况下,爱零食面临更大的融资压力与市场份额被挤压风险。截至 2025 年 6 月,其门店数虽增至 2000 家,但与行业头部品牌如鸣鸣很忙的 1.5 万家、万辰集团的近万家相比,差距悬殊。
2025 年 6 月 16 日,三只松鼠发布公告,宣布终止对湖南爱零食科技有限公司的收购计划。回溯至 2024 年 10 月,三只松鼠全资子公司安徽一件事创业投资有限公司拟斥资不超过 2 亿元,收购爱零食的控制权或相关业务及资产。彼时,双方签署的《投资意向协议》有效期为 90 天,并在今年 1 月将排他期延长至正式签约日,然而最终因部分核心条款未能达成一致,交易宣告流产。 野心勃勃的收购计划:三只松鼠的线下突围战 三只松鼠曾对爱零食寄予厚望,这场收购背后是其迫切的战略需求。从渠道结构来看,长期以来,三只松鼠线上依赖程度颇高。以 2024 年数据为例,其线上销售额占营业收入的比重为 69.73%,达 74.07 亿元 ,尽管同比增长 49.6%,但线下渠道的相对薄弱仍是制约其进一步发展的关键因素。 反观爱零食,作为量贩零食区域龙头,在湖南等多个区域已布局超 1800 家门店。在中国零食市场,2023 年线下销售占据了 82.6% 的市场份额 ,这一庞大的线下市场份额对三只松鼠而言极具吸引力。收购爱零食,无疑能让三只松鼠快速切入线下量贩零食市场,拓展线下版图,优化自身渠道结构,进而提升市场覆盖面和销售规模。 与此同时,量贩零食赛道正上演疯狂扩张。2024 年,零食很忙与赵一鸣零食合并为“鸣鸣很忙”,门店总数迅速突破 10000 家;万辰集团整合好想来、来优品等四大品牌,2025 年签约门店更是突破 1.5 万家。 爱零食在 2024 年 9 月门店数已超 1800 家,并提出“3 年 3000 家店” 的扩张目标,位居行业第五。在这样的竞争环境下,三只松鼠试图通过收购爱零食,在这场竞速赛中抢占一席之地,并借助双方在供应链、品牌上的互补优势,实现资源整合与产品矩阵升级。 财报数据:增长放缓下的战略权衡 三只松鼠 2024 年财报成绩亮眼:营收 106.22 亿元,同比增长 49.30%;归母净利润 4.08 亿元,同比增长 85.51%。线下分销收入也因并购爱折扣增长 48.62% 。但进入 2025 年第一季度,增速骤然放缓,营收仅微增 2.13%,归母净利润下滑 22.46%。 在此背景下,三只松鼠可能重新审视收购爱零食的必要性与可行性。或许是考虑到收购后整合的复杂性与成本,包括供应链整合、门店运营模式融合、人员管理等多方面潜在成本;以及自身业务调整与资源分配的需要。在面对与爱零食核心条款分歧时,选择终止收购。 命运转折:收购终止后的两家企业何去何从? 对于三只松鼠而言,终止收购虽短期内延缓了其在量贩零食赛道通过并购快速扩张的步伐,但也避免了潜在的整合风险与资金投入压力。公司可将资源集中于自有分销网络建设、社区零食店发展以及多品牌孵化等战略布局上。长期来看,如果三只松鼠能有效推进自身线下策略,持续提升供应链优势,进一步扩大线下分销规模,仍有望在零食市场竞争中巩固地位。 对爱零食来说,失去三只松鼠的资金注入与资源协同,门店扩张计划可能受阻。原本计划借助三只松鼠的 2 亿元投资加速门店扩张,目标在 2025 年达到 5000 家店,如今这一计划面临变数。在行业头部品牌竞争加剧的情况下,爱零食面临更大的融资压力与市场份额被挤压风险。截至 2025 年 6 月,其门店数虽增至 2000 家,但与行业头部品牌如鸣鸣很忙的 1.5 万家、万辰集团的近万家相比,差距悬殊。
考点1:”控制权”应该译为“controlling stake”,这是公司并购中的法律和商业术语。 考点2:“制约推荐”译为“bottleneck”。 考点3:”量贩零食“应该译为“snack discounter(s)”,量贩源自日语,指大量销售、薄利多销的模式,不可译成“mass-selling snacks”。 考点4:”万辰”应翻译为“Wanchen Group“ 考点5:“湖南”应翻译为”Hunan province“,表明其为省份 考点6:“爱折扣”音译或完全意译,不可两者结合 ,如“Ai Discount”
26905
新闻资讯
观点评论
34
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: 2. Theoretical BackgroundIn this section, we review the consumer acceptance literature. We focus on studies that examine willingness to buy, primarily in Germany, Europe and other Western countries, where data are available.2.1. Consumer AcceptanceBefore the determinants of consumer acceptance can be discussed, the concept of acceptance itself needs to be defined. The term “consumer acceptance” enjoys widespread usage across various research fields and has been extensively reviewed within the food science sector [28,29,30] but there remains a lack of consensus regarding its essential components and a general definition. This overarching construct is employed with ambiguity across different disciplines, including profitability and sufficient user adoption in the form of purchase intention, as well as the absence of categorical rejection towards a product or technology, indicative of a general willingness to engage with it [31,32]. In their scoping review, Baker et al. (2022) define consumer acceptance based on various outcome measurements such as general acceptance, willingness to buy, willingness to pay, willingness to try, perceptions, consumption or purchase intention [9]. Wassmann et al. (2021) include the following measures as correlates of willingness to consume in their meta-analysis: willingness to adopt, willingness to eat, willingness to pay, and willingness to try [10]. However, achieving a concise yet comprehensive definition remains a desirable objective, and such a task exceeds the scope of this paper.Nonetheless, it is crucial to define acceptance in each study to allow comparability. The existing literature on alternative protein sources centres around two main indicators, willingness to try [11,25,33,34] and willingness to buy [5,23,35,36,37,38,39,40], both of which are pivotal for gauging overall acceptance. Due to the lack of general guidance for selecting the indicators and given our specific interest in the potential of these three alternative protein sources to make substantial contributions to a change in diet, we have chosen to focus solely on willingness to buy, as it offers a direct insight into whether consumers are willing to make a purchase to include a specific protein source in their diets. Willingness to buy indicates the readiness to purchase food products of alternative protein sources (algae, crickets, jellyfish) in a grocery store, which goes beyond just the willingness to try [39]. In the following sections, we focus on the willingness to buy but also discuss the determinants of willingness to try and further correlates of consumer acceptance where previous studies on willingness to buy are lacking.2.2. NeophobiaFood neophobia is one of the most investigated determinants of consumer acceptance, particularly in the context of food innovations. Food neophobia, defined as the aversion to and dislike of novel foods, is a trait that manifests with varying degrees of intensity among individuals [41]. Food neophobia negatively influenced consumer acceptance of algae-based foods [42] as well as post-tasting satisfaction with algae-based food products [43]. Food neophobia had a negative impact on the adoption intention of Spirulina-enhanced foods (a specific type of microalgae) in a sub-sample of Belgian consumers who know a lot about the latest food trends and like to eat but not in the sub-samples of vegetarian and sporty Belgian consumers [44]. Food neophobia also showed a negative effect on the willingness to try [33] and to eat or consume insects among German consumers [45,46,47]. A similar effect could be found for the willingness to adopt insect-based foods [25] and to buy them [39]. A cross-country comparative study (involving the Czech Republic, Germany, Finland, and Sweden) further revealed a negative effect on the willingness to buy insect-based foods [38], while the meta-analysis indicated a negative effect on the willingness to consume [10]. Furthermore, food neophobia negatively influenced the attitude towards consuming jellyfish [40].When food is processed before sale, which is likely for the protein sources under study, a phobia against novel food technologies may also be relevant for consumer acceptance of such foods. Food technology neophobia is defined as the resistance to new or modern food technologies [48]. The effect on the consumer acceptance of algae-based products is not clear yet. Embling et al. (2022) did not find a significant effect [42]. However, the consumer willingness to consume insects [10,46] and readiness to adopt insects as meat substitutes [25] are influenced negatively by food technology neophobia. The relationship of food technology neophobia and consumer acceptance of jellyfish is still unclear.To summarise, it is reasonable to assume that food neophobia is associated with low consumer acceptance of alternative protein sources. However, the evident lack of understanding of how food technology neophobia impacts consumer acceptance of algae and jellyfish necessitates further investigations.2.3. Familiarity with New FoodsFamiliarity with the concept of consuming algae and insects as well as past consumption experiences are also important determinants of consumer acceptance [5,25,47]. Perceiving algae-based food as tasty, edible and familiar significantly increased acceptance [42]. People who are familiar with eating insects are more likely to adopt insects in their diets [25] and be willing to consume insects [10]. In contrast, Lammers et al. (2019) have found no effect of familiarity on willingness to consume insect burgers [46]. To the best of the authors’ knowledge, no studies have specifically investigated the impact of familiarity on the consumer acceptance of jellyfish-based foods. However, there appears to be a general tendency for familiarity with the sea environment to positively influence the consumption of jellyfish [40].Previous experiences with algae-based food have a positive impact on willingness to try and adopt algae into the diet [11]. Furthermore, experiences with eating insects lead to a higher willingness to try and adopt insect-based foods [11,33], to eat insect-based foods [39,45,46] and to buy insect-based food products [35,38]. However, a meta-analysis showed only a middle composite effect size of experience with willingness to consume insect-based foods (r = 0.35) [10]. Previous experiences with eating jellyfish lead to a higher willingness to try and adopt jellyfish in the diet [11].Based on the previous findings, both familiarity and experience might have a positive effect on consumer acceptance of algae, crickets and jellyfish. However, some clarity is needed considering scholars that have found no or only a small effect, necessitating a unified research approach.2.4. Food-Related HabitsFood-related lifestyle links individuals’ values with their eating behaviours, incorporating aspects such as food involvement and food innovativeness, as outlined in the food-related lifestyle framework by Brunsø et al. (2021) [49]. Food involvement refers to the extent to which individuals engage with and take an interest in different aspects of food [49]. Those with high food involvement regard food as an important part of their lives, dedicate considerable time and resources to it, and view it as essential for achieving personal goals [49,50]. Previous studies indicate that food involvement positively influences the intention to adopt Spirulina-enhanced foods among consumers who enjoy eating, though this effect is not observed among vegetarians, sporty individuals, or those well informed about the latest food trends [44]. Moreover, food involvement does not impact the willingness to try insect-based foods [51]. To the best of the authors’ knowledge, no studies have investigated the relationship between food involvement and the acceptance of jellyfish-based foods.Another aspect of the food-related lifestyle is food innovativeness, which refers to the extent to which individuals adopt and integrate novel culinary practices and ingredients into their dietary habits [49]. In the absence of previous research on the relationship of food innovativeness and the consumer acceptance of alternative protein sources, Mazhar & Zilahy (2023) show that food-related lifestyles (including food involvement and food innovativeness) are moderately correlated to green food purchasing behaviour [52]. Furthermore, sensation seeking, a related construct, has a positive effect on willingness to consume insects in Germany [46]. Moreover, a meta-analysis revealed that food sensation and innovation seeking have a strong composite effect on willingness to consume insects (r = 0.29) [10].A consistent research approach is required to bridge the knowledge gap about the direction and effect of food involvement and food innovativeness on consumer acceptance of algae, crickets and jellyfish.The meat consumption of consumers is an important determinant of consumer acceptance of proteins from alternative sources [5]. A strong belief in the benefits of eating meat is associated with lower acceptance of algae-based foods among meat consumers [42]. Additionally, consumers with a high rate of weekly meat intake are less likely to choose snacks made out of algae (besides lentils/beans snacks, insect snacks, and hybrid meat snacks), with both the number of meat-eating days and preferred portion size serving as significant predictors of their dietary preferences [53]. Those following a vegetarian or vegan diet are less willing to eat insects [47]. Furthermore, consumers who perceive meat as nutritious, healthy, and strongly focusing on its taste are less inclined to adopt insects as meat substitutes [25]. However, consumers intending to reduce their meat consumption show higher readiness to adopt insects as meat substitutes [25]. Contrary to these findings, Lammers et al. (2019) reported that higher meat consumption is linked to higher acceptance of food containing invisible insects, although this effect disappears when other factors influencing the acceptance (e.g., sensation seeking, sustainability consciousness, food disgust) are considered [46]. A meta-analysis found only a small effect size between willingness to consume insects and meat consumption (r = 0.08) [10]. Schäufele et al. (2019) found no effect of meat consumption on the willingness to try insect-based food [33]. The relationship between meat consumption and consumer acceptance of jellyfish is still unclear.In summary, it could be assumed that (high) meat consumption is associated with low consumer acceptance of algae-based protein. However, there is a clear knowledge gap regarding how meat consumption influences consumer acceptance of crickets and jellyfish. A consistent research approach is necessary to better understand its role.
2. Theoretical BackgroundIn this section, we review the consumer acceptance literature. We focus on studies that examine willingness to buy, primarily in Germany, Europe and other Western countries, where data are available.2.1. Consumer AcceptanceBefore the determinants of consumer acceptance can be discussed, the concept of acceptance itself needs to be defined. The term “consumer acceptance” enjoys widespread usage across various research fields and has been extensively reviewed within the food science sector [28,29,30] but there remains a lack of consensus regarding its essential components and a general definition. This overarching construct is employed with ambiguity across different disciplines, including profitability and sufficient user adoption in the form of purchase intention, as well as the absence of categorical rejection towards a product or technology, indicative of a general willingness to engage with it [31,32]. In their scoping review, Baker et al. (2022) define consumer acceptance based on various outcome measurements such as general acceptance, willingness to buy, willingness to pay, willingness to try, perceptions, consumption or purchase intention [9]. Wassmann et al. (2021) include the following measures as correlates of willingness to consume in their meta-analysis: willingness to adopt, willingness to eat, willingness to pay, and willingness to try [10]. However, achieving a concise yet comprehensive definition remains a desirable objective, and such a task exceeds the scope of this paper.Nonetheless, it is crucial to define acceptance in each study to allow comparability. The existing literature on alternative protein sources centres around two main indicators, willingness to try [11,25,33,34] and willingness to buy [5,23,35,36,37,38,39,40], both of which are pivotal for gauging overall acceptance. Due to the lack of general guidance for selecting the indicators and given our specific interest in the potential of these three alternative protein sources to make substantial contributions to a change in diet, we have chosen to focus solely on willingness to buy, as it offers a direct insight into whether consumers are willing to make a purchase to include a specific protein source in their diets. Willingness to buy indicates the readiness to purchase food products of alternative protein sources (algae, crickets, jellyfish) in a grocery store, which goes beyond just the willingness to try [39]. In the following sections, we focus on the willingness to buy but also discuss the determinants of willingness to try and further correlates of consumer acceptance where previous studies on willingness to buy are lacking.2.2. NeophobiaFood neophobia is one of the most investigated determinants of consumer acceptance, particularly in the context of food innovations. Food neophobia, defined as the aversion to and dislike of novel foods, is a trait that manifests with varying degrees of intensity among individuals [41]. Food neophobia negatively influenced consumer acceptance of algae-based foods [42] as well as post-tasting satisfaction with algae-based food products [43]. Food neophobia had a negative impact on the adoption intention of Spirulina-enhanced foods (a specific type of microalgae) in a sub-sample of Belgian consumers who know a lot about the latest food trends and like to eat but not in the sub-samples of vegetarian and sporty Belgian consumers [44]. Food neophobia also showed a negative effect on the willingness to try [33] and to eat or consume insects among German consumers [45,46,47]. A similar effect could be found for the willingness to adopt insect-based foods [25] and to buy them [39]. A cross-country comparative study (involving the Czech Republic, Germany, Finland, and Sweden) further revealed a negative effect on the willingness to buy insect-based foods [38], while the meta-analysis indicated a negative effect on the willingness to consume [10]. Furthermore, food neophobia negatively influenced the attitude towards consuming jellyfish [40].When food is processed before sale, which is likely for the protein sources under study, a phobia against novel food technologies may also be relevant for consumer acceptance of such foods. Food technology neophobia is defined as the resistance to new or modern food technologies [48]. The effect on the consumer acceptance of algae-based products is not clear yet. Embling et al. (2022) did not find a significant effect [42]. However, the consumer willingness to consume insects [10,46] and readiness to adopt insects as meat substitutes [25] are influenced negatively by food technology neophobia. The relationship of food technology neophobia and consumer acceptance of jellyfish is still unclear.To summarise, it is reasonable to assume that food neophobia is associated with low consumer acceptance of alternative protein sources. However, the evident lack of understanding of how food technology neophobia impacts consumer acceptance of algae and jellyfish necessitates further investigations.2.3. Familiarity with New FoodsFamiliarity with the concept of consuming algae and insects as well as past consumption experiences are also important determinants of consumer acceptance [5,25,47]. Perceiving algae-based food as tasty, edible and familiar significantly increased acceptance [42]. People who are familiar with eating insects are more likely to adopt insects in their diets [25] and be willing to consume insects [10]. In contrast, Lammers et al. (2019) have found no effect of familiarity on willingness to consume insect burgers [46]. To the best of the authors’ knowledge, no studies have specifically investigated the impact of familiarity on the consumer acceptance of jellyfish-based foods. However, there appears to be a general tendency for familiarity with the sea environment to positively influence the consumption of jellyfish [40].Previous experiences with algae-based food have a positive impact on willingness to try and adopt algae into the diet [11]. Furthermore, experiences with eating insects lead to a higher willingness to try and adopt insect-based foods [11,33], to eat insect-based foods [39,45,46] and to buy insect-based food products [35,38]. However, a meta-analysis showed only a middle composite effect size of experience with willingness to consume insect-based foods (r = 0.35) [10]. Previous experiences with eating jellyfish lead to a higher willingness to try and adopt jellyfish in the diet [11].Based on the previous findings, both familiarity and experience might have a positive effect on consumer acceptance of algae, crickets and jellyfish. However, some clarity is needed considering scholars that have found no or only a small effect, necessitating a unified research approach.2.4. Food-Related HabitsFood-related lifestyle links individuals’ values with their eating behaviours, incorporating aspects such as food involvement and food innovativeness, as outlined in the food-related lifestyle framework by Brunsø et al. (2021) [49]. Food involvement refers to the extent to which individuals engage with and take an interest in different aspects of food [49]. Those with high food involvement regard food as an important part of their lives, dedicate considerable time and resources to it, and view it as essential for achieving personal goals [49,50]. Previous studies indicate that food involvement positively influences the intention to adopt Spirulina-enhanced foods among consumers who enjoy eating, though this effect is not observed among vegetarians, sporty individuals, or those well informed about the latest food trends [44]. Moreover, food involvement does not impact the willingness to try insect-based foods [51]. To the best of the authors’ knowledge, no studies have investigated the relationship between food involvement and the acceptance of jellyfish-based foods.Another aspect of the food-related lifestyle is food innovativeness, which refers to the extent to which individuals adopt and integrate novel culinary practices and ingredients into their dietary habits [49]. In the absence of previous research on the relationship of food innovativeness and the consumer acceptance of alternative protein sources, Mazhar & Zilahy (2023) show that food-related lifestyles (including food involvement and food innovativeness) are moderately correlated to green food purchasing behaviour [52]. Furthermore, sensation seeking, a related construct, has a positive effect on willingness to consume insects in Germany [46]. Moreover, a meta-analysis revealed that food sensation and innovation seeking have a strong composite effect on willingness to consume insects (r = 0.29) [10].A consistent research approach is required to bridge the knowledge gap about the direction and effect of food involvement and food innovativeness on consumer acceptance of algae, crickets and jellyfish.The meat consumption of consumers is an important determinant of consumer acceptance of proteins from alternative sources [5]. A strong belief in the benefits of eating meat is associated with lower acceptance of algae-based foods among meat consumers [42]. Additionally, consumers with a high rate of weekly meat intake are less likely to choose snacks made out of algae (besides lentils/beans snacks, insect snacks, and hybrid meat snacks), with both the number of meat-eating days and preferred portion size serving as significant predictors of their dietary preferences [53]. Those following a vegetarian or vegan diet are less willing to eat insects [47]. Furthermore, consumers who perceive meat as nutritious, healthy, and strongly focusing on its taste are less inclined to adopt insects as meat substitutes [25]. However, consumers intending to reduce their meat consumption show higher readiness to adopt insects as meat substitutes [25]. Contrary to these findings, Lammers et al. (2019) reported that higher meat consumption is linked to higher acceptance of food containing invisible insects, although this effect disappears when other factors influencing the acceptance (e.g., sensation seeking, sustainability consciousness, food disgust) are considered [46]. A meta-analysis found only a small effect size between willingness to consume insects and meat consumption (r = 0.08) [10]. Schäufele et al. (2019) found no effect of meat consumption on the willingness to try insect-based food [33]. The relationship between meat consumption and consumer acceptance of jellyfish is still unclear.In summary, it could be assumed that (high) meat consumption is associated with low consumer acceptance of algae-based protein. However, there is a clear knowledge gap regarding how meat consumption influences consumer acceptance of crickets and jellyfish. A consistent research approach is necessary to better understand its role.
考点1:【desirable objective】应该译为“理想目标”。 考点2:【existing literature】应译为“现有文献”。
26ab2
学术论文
社会科学
63
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: As the dominant discourse would have it, the economic world is a pure and perfect order, implacably unrolling the logic of its predictable consequences, and prompt to repress all violations by the sanctions that it inflicts, either automatically or - more unusually - through the intermediary of its armed extensions, the International Monetary Fund (IMF) and the Organization for Economic Cooperation and Development (OECD) and the policies they impose: reducing labour costs, reducing public expenditures and making work more flexible. Is the dominant discourse right? What if, in reality, this economic order were no more than the implementation of a utopia - the utopia of neoliberalism - thus converted into a political problem? One that, with the aid of the economic theory that it proclaims, succeeds in conceiving of itself as the scientific description of reality? This tutelary theory is a pure mathematical fiction. From the start it has been founded on a formidable abstraction. For, in the name of a narrow and strict conception of rationality as individual rationality, it brackets the economic and social conditions of rational orientations and the economic and social structures that are the condition of their application. To give the measure of this omission, it is enough to think just of the educational system. Education is never taken account of as such at a time when it plays a determining role in the production of goods and services as in the production of the producers themselves. From this sort of original sin, inscribed in the Walrasian myth (1) of "pure theory", flow all of the deficiencies and faults of the discipline of economics and the fatal obstinacy with which it attaches itself to the arbitrary opposition which it induces, through its mere existence, between a properly economic logic, based on competition and efficiency, and social logic, which is subject to the rule of fairness. That said, this "theory" that is desocialised and dehistoricised at its roots has, today more than ever, the means of making itself true and empirically verifiable. In effect, neoliberal discourse is not just one discourse among many. Rather, it is a "strong discourse" - the way psychiatric discourse is in an asylum, in Erving Goffman's analysis (2). It is so strong and so hard to combat only because it has on its side all of the forces of a world of relations of forces, a world that it contributes to making what it is. It does this most notably by orienting the economic choices of those who dominate economic relationships. It thus adds its own symbolic force to these relations of forces. In the name of this scientific programme, converted into a plan of political action, an immense political project is underway, although its status as such is denied because it appears to be purely negative. This project aims to create the conditions under which the "theory" can be realised and can function: a programme of the methodical destruction of collectives. The movement toward the neoliberal utopia of a pure and perfect market is made possible by the politics of financial deregulation. And it is achieved through the transformative and, it must be said, destructive action of all of the political measures (of which the most recent is the Multilateral Agreement on Investment (MAI), designed to protect foreign corporations and their investments from national states) that aim to call into question any and all collective structures that could serve as an obstacle to the logic of the pure market: the nation, whose space to manoeuvre continually decreases; work groups, for example through the individualisation of salaries and of careers as a function of individual competences, with the consequent atomisation of workers; collectives for the defence of the rights of workers, unions, associations, cooperatives; even the family, which loses part of its control over consumption through the constitution of markets by age groups. The neoliberal programme draws its social power from the political and economic power of those whose interests it expresses: stockholders, financial operators, industrialists, conservative or social-democratic politicians who have been converted to the reassuring layoffs of laisser-faire, high-level financial officials eager to impose policies advocating their own extinction because, unlike the managers of firms, they run no risk of having eventually to pay the consequences. Neoliberalism tends on the whole to favour severing the economy from social realities and thereby constructing, in reality, an economic system conforming to its description in pure theory, that is a sort of logical machine that presents itself as a chain of constraints regulating economic agents. The globalisation of financial markets, when joined with the progress of information technology, ensures an unprecedented mobility of capital. It gives investors concerned with the short-term profitability of their investments the possibility of permanently comparing the profitability of the largest corporations and, in consequence, penalising these firms' relative setbacks. Subjected to this permanent threat, the corporations themselves have to adjust more and more rapidly to the exigencies of the markets, under penalty of "losing the market's confidence", as they say, as well as the support of their stockholders. The latter, anxious to obtain short-term profits, are more and more able to impose their will on managers, using financial directorates to establish the rules under which managers operate and to shape their policies regarding hiring, employment, and wages. Thus the absolute reign of flexibility is established, with employees being hiring on fixed-term contracts or on a temporary basis and repeated corporate restructurings and, within the firm itself, competition among autonomous divisions as well as among teams forced to perform multiple functions. Finally, this competition is extended to individuals themselves, through the individualisation of the wage relationship: establishment of individual performance objectives, individual performance evaluations, permanent evaluation, individual salary increases or granting of bonuses as a function of competence and of individual merit; individualised career paths; strategies of "delegating responsibility" tending to ensure the self-exploitation of staff who, simple wage labourers in relations of strong hierarchical dependence, are at the same time held responsible for their sales, their products, their branch, their store, etc. as though they were independent contractors. This pressure toward "self-control" extends workers' "involvement" according to the techniques of "participative management" considerably beyond management level. All of these are techniques of rational domination that impose over-involvement in work (and not only among management) and work under emergency or high-stress conditions. And they converge to weaken or abolish collective standards or solidarities (3). In this way, a Darwinian world emerges - it is the struggle of all against all at all levels of the hierarchy, which finds support through everyone clinging to their job and organisation under conditions of insecurity, suffering, and stress. Without a doubt, the practical establishment of this world of struggle would not succeed so completely without the complicity of all of the precarious arrangements that produce insecurity and of the existence of a reserve army of employees rendered docile by these social processes that make their situations precarious, as well as by the permanent threat of unemployment. This reserve army exists at all levels of the hierarchy, even at the higher levels, especially among managers. The ultimate foundation of this entire economic order placed under the sign of freedom is in effect the structural violence of unemployment, of the insecurity of job tenure and the menace of layoff that it implies. The condition of the "harmonious" functioning of the individualist micro-economic model is a mass phenomenon, the existence of a reserve army of the unemployed. This structural violence also weighs on what is called the labour contract (wisely rationalised and rendered unreal by the "theory of contracts"). Organisational discourse has never talked as much of trust, co-operation, loyalty, and organisational culture as in an era when adherence to the organisation is obtained at each moment by eliminating all temporal guarantees of employment (three-quarters of hires are for fixed duration, the proportion of temporary employees keeps rising, employment "at will" and the right to fire an individual tend to be freed from any restriction).
As the dominant discourse would have it, the economic world is a pure and perfect order, implacably unrolling the logic of its predictable consequences, and prompt to repress all violations by the sanctions that it inflicts, either automatically or - more unusually - through the intermediary of its armed extensions, the International Monetary Fund (IMF) and the Organization for Economic Cooperation and Development (OECD) and the policies they impose: reducing labour costs, reducing public expenditures and making work more flexible. Is the dominant discourse right? What if, in reality, this economic order were no more than the implementation of a utopia - the utopia of neoliberalism - thus converted into a political problem? One that, with the aid of the economic theory that it proclaims, succeeds in conceiving of itself as the scientific description of reality? This tutelary theory is a pure mathematical fiction. From the start it has been founded on a formidable abstraction. For, in the name of a narrow and strict conception of rationality as individual rationality, it brackets the economic and social conditions of rational orientations and the economic and social structures that are the condition of their application. To give the measure of this omission, it is enough to think just of the educational system. Education is never taken account of as such at a time when it plays a determining role in the production of goods and services as in the production of the producers themselves. From this sort of original sin, inscribed in the Walrasian myth (1) of "pure theory", flow all of the deficiencies and faults of the discipline of economics and the fatal obstinacy with which it attaches itself to the arbitrary opposition which it induces, through its mere existence, between a properly economic logic, based on competition and efficiency, and social logic, which is subject to the rule of fairness. That said, this "theory" that is desocialised and dehistoricised at its roots has, today more than ever, the means of making itself true and empirically verifiable. In effect, neoliberal discourse is not just one discourse among many. Rather, it is a "strong discourse" - the way psychiatric discourse is in an asylum, in Erving Goffman's analysis (2). It is so strong and so hard to combat only because it has on its side all of the forces of a world of relations of forces, a world that it contributes to making what it is. It does this most notably by orienting the economic choices of those who dominate economic relationships. It thus adds its own symbolic force to these relations of forces. In the name of this scientific programme, converted into a plan of political action, an immense political project is underway, although its status as such is denied because it appears to be purely negative. This project aims to create the conditions under which the "theory" can be realised and can function: a programme of the methodical destruction of collectives. The movement toward the neoliberal utopia of a pure and perfect market is made possible by the politics of financial deregulation. And it is achieved through the transformative and, it must be said, destructive action of all of the political measures (of which the most recent is the Multilateral Agreement on Investment (MAI), designed to protect foreign corporations and their investments from national states) that aim to call into question any and all collective structures that could serve as an obstacle to the logic of the pure market: the nation, whose space to manoeuvre continually decreases; work groups, for example through the individualisation of salaries and of careers as a function of individual competences, with the consequent atomisation of workers; collectives for the defence of the rights of workers, unions, associations, cooperatives; even the family, which loses part of its control over consumption through the constitution of markets by age groups. The neoliberal programme draws its social power from the political and economic power of those whose interests it expresses: stockholders, financial operators, industrialists, conservative or social-democratic politicians who have been converted to the reassuring layoffs of laisser-faire, high-level financial officials eager to impose policies advocating their own extinction because, unlike the managers of firms, they run no risk of having eventually to pay the consequences. Neoliberalism tends on the whole to favour severing the economy from social realities and thereby constructing, in reality, an economic system conforming to its description in pure theory, that is a sort of logical machine that presents itself as a chain of constraints regulating economic agents. The globalisation of financial markets, when joined with the progress of information technology, ensures an unprecedented mobility of capital. It gives investors concerned with the short-term profitability of their investments the possibility of permanently comparing the profitability of the largest corporations and, in consequence, penalising these firms' relative setbacks. Subjected to this permanent threat, the corporations themselves have to adjust more and more rapidly to the exigencies of the markets, under penalty of "losing the market's confidence", as they say, as well as the support of their stockholders. The latter, anxious to obtain short-term profits, are more and more able to impose their will on managers, using financial directorates to establish the rules under which managers operate and to shape their policies regarding hiring, employment, and wages. Thus the absolute reign of flexibility is established, with employees being hiring on fixed-term contracts or on a temporary basis and repeated corporate restructurings and, within the firm itself, competition among autonomous divisions as well as among teams forced to perform multiple functions. Finally, this competition is extended to individuals themselves, through the individualisation of the wage relationship: establishment of individual performance objectives, individual performance evaluations, permanent evaluation, individual salary increases or granting of bonuses as a function of competence and of individual merit; individualised career paths; strategies of "delegating responsibility" tending to ensure the self-exploitation of staff who, simple wage labourers in relations of strong hierarchical dependence, are at the same time held responsible for their sales, their products, their branch, their store, etc. as though they were independent contractors. This pressure toward "self-control" extends workers' "involvement" according to the techniques of "participative management" considerably beyond management level. All of these are techniques of rational domination that impose over-involvement in work (and not only among management) and work under emergency or high-stress conditions. And they converge to weaken or abolish collective standards or solidarities (3). In this way, a Darwinian world emerges - it is the struggle of all against all at all levels of the hierarchy, which finds support through everyone clinging to their job and organisation under conditions of insecurity, suffering, and stress. Without a doubt, the practical establishment of this world of struggle would not succeed so completely without the complicity of all of the precarious arrangements that produce insecurity and of the existence of a reserve army of employees rendered docile by these social processes that make their situations precarious, as well as by the permanent threat of unemployment. This reserve army exists at all levels of the hierarchy, even at the higher levels, especially among managers. The ultimate foundation of this entire economic order placed under the sign of freedom is in effect the structural violence of unemployment, of the insecurity of job tenure and the menace of layoff that it implies. The condition of the "harmonious" functioning of the individualist micro-economic model is a mass phenomenon, the existence of a reserve army of the unemployed. This structural violence also weighs on what is called the labour contract (wisely rationalised and rendered unreal by the "theory of contracts"). Organisational discourse has never talked as much of trust, co-operation, loyalty, and organisational culture as in an era when adherence to the organisation is obtained at each moment by eliminating all temporal guarantees of employment (three-quarters of hires are for fixed duration, the proportion of temporary employees keeps rising, employment "at will" and the right to fire an individual tend to be freed from any restriction).
考点1:“ the dominant discourse”译为“主流话语;主导性话语” 考点2:“armed extensions”需注意译文措辞与前后语句的匹配度 考点3:“tutelary theory”推荐译为“指导性理论;保护性理论” 考点4: “bracket”需译为 “悬置;搁置;存而不论” 考点5: “strong discourse”应译为“强势话语” 考点6:“the struggle of all against all”应译为“所有人对所有人的斗争/战争” 考点7:“ reserve army of employees”必须译为“劳动后备军”,不可译为其他非政经术语 考点8:“the theory of contracts”应译为“契约理论”
26eab
学术论文
社会科学
4
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Results: Evaluation of FS All the published guidelines agree that no further investigation is routinely necessary for children with a single episode of SFS[8]-[14] (Table 2). The British guidelines suggest measurement of blood glucose concentration if the child is still convulsing or unrousable when seen either with or without fever and to consider neuroimaging before lumbar puncture (LP) if the child is comatose.[10] According to ILAE guidelines, it is recommended to search for the fever’s etiology in the case of CFS. Additionally, neuroimaging is recommended to investigate a possible underlying brain lesion. Blood chemical tests are suggested in relation to specific clinical conditions.[8] The AAP underscores the importance of evaluating the underlying cause of fever but recommends that blood tests should not be performed routinely for the sole purpose of identifying the cause of SFS, since there is no evidence to suggest that routine blood studies are of benefit in the evaluation of the child with SFS and there could be a preponderance of harm over benefit.[12] Japanese guidelines suggest that blood examination should be considered in cases of poor general condition, prolonged altered consciousness, or signs of dehydration.[9] These guidelines also recommend neuroimaging only before LP, as previously stated in the British guidelines,[10] due to the risk of cerebral herniation. By contrast, the Mexican guidelines recommend neuroimaging in patients with FS who do not regain complete consciousness in hours, with prolonged Todd paralysis, or with other focal neurological alterations.[13] According to the AOCN guidelines, urine analysis should be considered in children aged <18 months, as most children do not have a focus of infection, and magnetic resonance imaging (MRI) should be considered in children with the first episode of CFS with prolonged or focal features (within 72 hours).[14] The role of the EEG In all the existing guidelines, electroencephalography (EEG) is not recommended in case of a first episode of SFS, due to its limited diagnostic value.[8,9,13,14] ILAE guidelines recommend performing an EEG in CFS, owing to the high diagnostic value it may have for some viral encephalitides.[8] Similarly, according to the Japanese guidelines, EEG may have a role in differentiating between acute encephalopathy and CFS, but few peer-reviewed reports have appeared in literature validating the utility of EEG for such differentiation.[9] Even though the Japanese guidelines conclude that EEG examination is not indicated for children with SFS, the authors highlight the possible correlation between epileptiform discharges on EEG and development of epilepsy, as well as its usefulness in the diagnosis of acute encephalopathy. Moreover, the Mexican guidelines recommended EEG in a focal and/or prolonged convulsion, or in CFS if the neurological examination shows alterations.[13] According to the AOCN, EEG should be considered in children with CFS or in children with focal findings on neuroimaging.[14] In cases of prolonged altered mental status, the EEG is not useful to predict recurrence of FS or future epilepsy. Moreover, it is important to note that if EEG were useful to identify viral encephalitis, that would establish an alternative diagnosis of viral encephalitis, causing both fever and provoked seizure, and potentially invalidate the initial working diagnosis of CFS. The AOCN suggests that EEG is not recommended in developmentally normal children with SFS because it does not predict the recurrence of FS or subsequent epilepsy.[14] The role of the LP All the published guidelines recommend performing LP if there are meningeal signs or if a CNS infection is suspected (see Table 2). According to the British guidelines, LP should also be performed after a CFS or if the child is unduly drowsy or irritable or systemically ill.[10] The guidelines also state that LP should probably be performed in children aged less than 18 months and almost certainly in children aged less than 12 months, regardless of meningeal signs.[10] They specify that ideally the decision should be taken by an experienced doctor, who may decide on clinical grounds that LP is unnecessary even in a younger child, but when in doubt the investigation should be performed.[11] Similarly, the AOCN guidelines declare that LP should be considered in children aged <12 months with first episode of FS and in children aged >12 months who have been pretreated with antibiotics.[14] The age threshold for performing LP outlined in these older recommendations differs from the most recent ones.[8,12,13] In particular, the AAP states that LP is an option for any infant aged between six and 12 months who has not received scheduled immunizations for Haemophilus influenzae type B or Streptococcus pneumoniae or when the immunization status cannot be determined.[12] Additionally, LP may be considered when the child has recently been treated with antibiotics, as suggested by clinical experience and case series. LP may be an option in the child who is pretreated with antibiotics, because antibiotic treatment can mask the signs and symptoms of meningitis.[12] Similarly, the Mexican guidelines state that LP is indicated in children aged <12 months who have not completed their immunizations or have received previous antibiotic treatment or in children aged less than six months with SFS unless an experienced pediatrician evaluates the patient before two hours. ILAE guidelines do not give mandatory indications on whether to routinely perform LP in younger patients, but suggest a careful observation of children aged <18 months for at least 24 hours, since at this age clinical signs and symptoms for meningitis may be minimal and LP must be performed in the presence of meningeal signs.[8,10,12] Moreover, they suggest considering LP also in patients under antibiotic treatment during the days before FS.[8] Pharmacologic prophylaxis and home treatment of FS Almost all published guidelines state that neither intermittent nor continuous seizure prophylaxis is recommended in SFS (see Table 2), except for few cases mentioned below.[8,11,13,14] The prevailing recommendation against prophylactic treatment stems from the consideration that such medications carry significant side effects and do not prevent the development of epilepsy. According to ILAE, intermittent prophylaxis may be considered in cases with high risk of recurrence, defined as at least three episodes in six months or at least four episodes in one year, or CFS due to prolonged duration lasting more than 15 minutes or requiring pharmacologic therapy to be stopped.[8,16–18] Intermittent prophylaxis of FS consists in administering rectal (first choice) or oral diazepam at the onset of fever, to be repeated if fever persists after eight hours.[8] In case of failure of intermittent prophylaxis and when parents are unable to promptly recognize the onset of fever, ILAE guidelines state that continuous ASMs with phenobarbital or valproic acid may be used.[8] as previously stated in the Japanese guidelines published in 1996.[19] The AAP states that neither continuous nor intermittent seizure prophylaxis is recommended for children with one or more SFS, based on randomized controlled trials and diagnostic studies with minor limitations.[11,16,17] Nonetheless, according to the AAP guidelines, intermittent oral diazepam at the onset of febrile illness may be considered and effective in preventing recurrence in cases of severe parental anxiety associated with FS.[11] Japanese guidelines suggest considering prophylaxis with ASM also in case of severe parental anxiety.[9] Japanese guidelines recommend the prophylactic use of diazepam during febrile illness in children at high risk of recurrence of FS.[9] As suggested in the study conducted by Berg et al., and Pavlidou et al., who both studied children with a first episode of FS,[18,20] risk factors of recurrence include young age at onset, positive family history of FS, abnormal perinatal history, low degree of fever while in the emergency department (ED), brief duration between the onset of fever and the initial seizure, recurrence within the same illness, partial onset, or focal features of FS. According to several studies, the use of prophylactic diazepam for the prevention of FS is indicated in children with a history of a prolonged FS (lasting longer than 15 minutes) and in children with repeated FS and two of the following risk factors: focal or repeated seizures within 24 hours, pre-existing neurological abnormality or developmental delay, family history of FS or epilepsy, age less than 12 months, seizure within one hour after the onset of fever, or seizure occurring with body temperature less than 38°C[9][21–24]. In the British guidelines long-term prophylaxis with ASM (valproic acid, phenobarbitone) is not recommended, according to previous studies,[25] except in case of frequent recurrences.[10] According to the Mexican guidelines, treatment with ASM should be considered in patients with long or repetitive FS despite the prophylactic use of diazepam.[13] The AOCN guidelines recommend considering intermittent prophylaxis among children with frequent recurrent SFS with parental anxiety, residence far from medical facilities, and those with CFS who have not been started on continuous prophylaxis. On the other hand, continuous prophylaxis should be considered among children with febrile status epilepticus, in children with neurodevelopmental delay, and in frequent CFS.[14] Home treatment for FS is recommended in four guidelines (UK, ILAE, Mexican, AOCN). An ASM (rectal diazepam or intranasal midazolam) should be administered if FS last more than three minutes (ILAE), five minutes (UK, Mexican), or three to five minutes (AOCN). In particular, ILAE recommends administering rectal diazepam only if the seizure lasts longer than three minutes.[8] In the British guidelines there is no agreement on the timing of administration of rectal diazepam: some members of the working group—although neither percentages of members nor studies on which the statements are based on are specified—suggest administering it as soon as possible after the onset of a convulsion, whereas others advised waiting for five minutes, by which time most convulsions will have stopped and ASMs will not be necessary.[10] Indeed, the use of therapies to prevent subsequent epilepsy is not recommended.[8] The British, Mexican, and AOCN guidelines underline the importance of parental education and suggest providing exhaustive information to parents both verbally and possibly written.[10,13,14] Health education should include a description of features of FS and instructions on how to manage possible recurrences.[8]
Results: Evaluation of FS All the published guidelines agree that no further investigation is routinely necessary for children with a single episode of SFS[8]-[14] (Table 2). The British guidelines suggest measurement of blood glucose concentration if the child is still convulsing or unrousable when seen either with or without fever and to consider neuroimaging before lumbar puncture (LP) if the child is comatose.[10] According to ILAE guidelines, it is recommended to search for the fever’s etiology in the case of CFS. Additionally, neuroimaging is recommended to investigate a possible underlying brain lesion. Blood chemical tests are suggested in relation to specific clinical conditions.[8] The AAP underscores the importance of evaluating the underlying cause of fever but recommends that blood tests should not be performed routinely for the sole purpose of identifying the cause of SFS, since there is no evidence to suggest that routine blood studies are of benefit in the evaluation of the child with SFS and there could be a preponderance of harm over benefit.[12] Japanese guidelines suggest that blood examination should be considered in cases of poor general condition, prolonged altered consciousness, or signs of dehydration.[9] These guidelines also recommend neuroimaging only before LP, as previously stated in the British guidelines,[10] due to the risk of cerebral herniation. By contrast, the Mexican guidelines recommend neuroimaging in patients with FS who do not regain complete consciousness in hours, with prolonged Todd paralysis, or with other focal neurological alterations.[13] According to the AOCN guidelines, urine analysis should be considered in children aged <18 months, as most children do not have a focus of infection, and magnetic resonance imaging (MRI) should be considered in children with the first episode of CFS with prolonged or focal features (within 72 hours).[14] The role of the EEG In all the existing guidelines, electroencephalography (EEG) is not recommended in case of a first episode of SFS, due to its limited diagnostic value.[8,9,13,14] ILAE guidelines recommend performing an EEG in CFS, owing to the high diagnostic value it may have for some viral encephalitides.[8] Similarly, according to the Japanese guidelines, EEG may have a role in differentiating between acute encephalopathy and CFS, but few peer-reviewed reports have appeared in literature validating the utility of EEG for such differentiation.[9] Even though the Japanese guidelines conclude that EEG examination is not indicated for children with SFS, the authors highlight the possible correlation between epileptiform discharges on EEG and development of epilepsy, as well as its usefulness in the diagnosis of acute encephalopathy. Moreover, the Mexican guidelines recommended EEG in a focal and/or prolonged convulsion, or in CFS if the neurological examination shows alterations.[13] According to the AOCN, EEG should be considered in children with CFS or in children with focal findings on neuroimaging.[14] In cases of prolonged altered mental status, the EEG is not useful to predict recurrence of FS or future epilepsy. Moreover, it is important to note that if EEG were useful to identify viral encephalitis, that would establish an alternative diagnosis of viral encephalitis, causing both fever and provoked seizure, and potentially invalidate the initial working diagnosis of CFS. The AOCN suggests that EEG is not recommended in developmentally normal children with SFS because it does not predict the recurrence of FS or subsequent epilepsy.[14] The role of the LP All the published guidelines recommend performing LP if there are meningeal signs or if a CNS infection is suspected (see Table 2). According to the British guidelines, LP should also be performed after a CFS or if the child is unduly drowsy or irritable or systemically ill.[10] The guidelines also state that LP should probably be performed in children aged less than 18 months and almost certainly in children aged less than 12 months, regardless of meningeal signs.[10] They specify that ideally the decision should be taken by an experienced doctor, who may decide on clinical grounds that LP is unnecessary even in a younger child, but when in doubt the investigation should be performed.[11] Similarly, the AOCN guidelines declare that LP should be considered in children aged <12 months with first episode of FS and in children aged >12 months who have been pretreated with antibiotics.[14] The age threshold for performing LP outlined in these older recommendations differs from the most recent ones.[8,12,13] In particular, the AAP states that LP is an option for any infant aged between six and 12 months who has not received scheduled immunizations for Haemophilus influenzae type B or Streptococcus pneumoniae or when the immunization status cannot be determined.[12] Additionally, LP may be considered when the child has recently been treated with antibiotics, as suggested by clinical experience and case series. LP may be an option in the child who is pretreated with antibiotics, because antibiotic treatment can mask the signs and symptoms of meningitis.[12] Similarly, the Mexican guidelines state that LP is indicated in children aged <12 months who have not completed their immunizations or have received previous antibiotic treatment or in children aged less than six months with SFS unless an experienced pediatrician evaluates the patient before two hours. ILAE guidelines do not give mandatory indications on whether to routinely perform LP in younger patients, but suggest a careful observation of children aged <18 months for at least 24 hours, since at this age clinical signs and symptoms for meningitis may be minimal and LP must be performed in the presence of meningeal signs.[8,10,12] Moreover, they suggest considering LP also in patients under antibiotic treatment during the days before FS.[8] Pharmacologic prophylaxis and home treatment of FS Almost all published guidelines state that neither intermittent nor continuous seizure prophylaxis is recommended in SFS (see Table 2), except for few cases mentioned below.[8,11,13,14] The prevailing recommendation against prophylactic treatment stems from the consideration that such medications carry significant side effects and do not prevent the development of epilepsy. According to ILAE, intermittent prophylaxis may be considered in cases with high risk of recurrence, defined as at least three episodes in six months or at least four episodes in one year, or CFS due to prolonged duration lasting more than 15 minutes or requiring pharmacologic therapy to be stopped.[8,16–18] Intermittent prophylaxis of FS consists in administering rectal (first choice) or oral diazepam at the onset of fever, to be repeated if fever persists after eight hours.[8] In case of failure of intermittent prophylaxis and when parents are unable to promptly recognize the onset of fever, ILAE guidelines state that continuous ASMs with phenobarbital or valproic acid may be used.[8] as previously stated in the Japanese guidelines published in 1996.[19] The AAP states that neither continuous nor intermittent seizure prophylaxis is recommended for children with one or more SFS, based on randomized controlled trials and diagnostic studies with minor limitations.[11,16,17] Nonetheless, according to the AAP guidelines, intermittent oral diazepam at the onset of febrile illness may be considered and effective in preventing recurrence in cases of severe parental anxiety associated with FS.[11] Japanese guidelines suggest considering prophylaxis with ASM also in case of severe parental anxiety.[9] Japanese guidelines recommend the prophylactic use of diazepam during febrile illness in children at high risk of recurrence of FS.[9] As suggested in the study conducted by Berg et al., and Pavlidou et al., who both studied children with a first episode of FS,[18,20] risk factors of recurrence include young age at onset, positive family history of FS, abnormal perinatal history, low degree of fever while in the emergency department (ED), brief duration between the onset of fever and the initial seizure, recurrence within the same illness, partial onset, or focal features of FS. According to several studies, the use of prophylactic diazepam for the prevention of FS is indicated in children with a history of a prolonged FS (lasting longer than 15 minutes) and in children with repeated FS and two of the following risk factors: focal or repeated seizures within 24 hours, pre-existing neurological abnormality or developmental delay, family history of FS or epilepsy, age less than 12 months, seizure within one hour after the onset of fever, or seizure occurring with body temperature less than 38°C[9][21–24]. In the British guidelines long-term prophylaxis with ASM (valproic acid, phenobarbitone) is not recommended, according to previous studies,[25] except in case of frequent recurrences.[10] According to the Mexican guidelines, treatment with ASM should be considered in patients with long or repetitive FS despite the prophylactic use of diazepam.[13] The AOCN guidelines recommend considering intermittent prophylaxis among children with frequent recurrent SFS with parental anxiety, residence far from medical facilities, and those with CFS who have not been started on continuous prophylaxis. On the other hand, continuous prophylaxis should be considered among children with febrile status epilepticus, in children with neurodevelopmental delay, and in frequent CFS.[14] Home treatment for FS is recommended in four guidelines (UK, ILAE, Mexican, AOCN). An ASM (rectal diazepam or intranasal midazolam) should be administered if FS last more than three minutes (ILAE), five minutes (UK, Mexican), or three to five minutes (AOCN). In particular, ILAE recommends administering rectal diazepam only if the seizure lasts longer than three minutes.[8] In the British guidelines there is no agreement on the timing of administration of rectal diazepam: some members of the working group—although neither percentages of members nor studies on which the statements are based on are specified—suggest administering it as soon as possible after the onset of a convulsion, whereas others advised waiting for five minutes, by which time most convulsions will have stopped and ASMs will not be necessary.[10] Indeed, the use of therapies to prevent subsequent epilepsy is not recommended.[8] The British, Mexican, and AOCN guidelines underline the importance of parental education and suggest providing exhaustive information to parents both verbally and possibly written.[10,13,14] Health education should include a description of features of FS and instructions on how to manage possible recurrences.[8]
考点1:“Febrile Seizures (FS)”推荐译为“热性惊厥”,【关键内容,为全文主题】 考点2:“Simple Febrile Seizures (SFS)”推荐译为“单纯性热性惊厥”, 考点3:“Complex Febrile Seizures (CFS)”推荐译为“复杂性热性惊厥”, 考点4:“Lumbar Puncture (LP)”推荐译为“腰椎穿刺”, 考点5:“Neuroimaging”推荐译为“神经影像学检查”, 考点6:“Electroencephalography (EEG)”推荐译为“脑电图”, 考点7:“Central Nervous System (CNS) Infection”推荐译为“中枢神经系统感染”, 考点8:“Cerebral Herniation”推荐译为“脑疝”, 考点9:“Todd Paralysis”推荐译为“托德瘫痪”, 考点10:“Focal Neurological Alterations”推荐译为“局灶性神经功能异常”,【易错点 "alterations"在神经科特指"功能异常",非泛指"系统异常"】 考点11:“Urine Analysis”推荐译为“尿液分析”, 考点12:“Magnetic Resonance Imaging (MRI)”推荐译为“磁共振成像”, 考点13:“Viral Encephalitides”推荐译为“病毒性脑炎”, 考点14:“Acute Encephalopathy”推荐译为“急性脑病”, 考点15:“Epileptiform Discharges”推荐译为“癫痫样放电”, 考点16:“Meningeal Signs”推荐译为“脑膜刺激征”, 考点17:“Immunization Status”推荐译为“免疫接种状态”, 考点18:“Haemophilus influenzae type B (HIB)”推荐译为“b型流感嗜血杆菌”, 考点19:“Streptococcus pneumoniae”推荐译为“肺炎链球菌”, 考点20:“Continuous Prophylaxis”推荐译为“持续性预防”,【易错点 "Continuous"在文中应翻译为持续性而不是连续或长期】 考点21:“Antiseizure Medications (ASMs)”推荐译为“抗癫痫药物”, 考点22:“within the same illness”推荐译为“同一病程内”,【易错点 结合上下文illness应翻译为"病程",而不是"疾病"。结合上下文illness应翻译为"病程",而不是"疾病"】 考点23:“and possibly written”推荐译为“书面材料”,【易错点 结合上下文written应翻译为书面材料,比如出院小结、告知书等材料,而不是信息、书信等】 考点24:“residence far from medical facilities”推荐译为“居住地远离医疗机构”,【"facilities"在文中语境应译"机构",比如医院等,设施指CT等设备,而不是"设施"】 考点25:“abnormal perinatal history”推荐译为“围产期异常史”,【易错点 医学专业表达习惯为"异常史",是指病史中的既往史部分,若翻译为"病史异常"则不专业且容易有歧义】 考点26:“AOCN (Association of Child Neurology India)”推荐译为“印度儿童神经病学协会”。
279a3
学术论文
自然科学
101
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Real estate CEO warns of growing ‘exodus’ as people have ‘given up’ on California — but where are they going? With its beautiful weather, breathtaking coastlines and vibrant culture, California has always held a special allure. But according to Don Peebles — founder, chairman and CEO of real estate investment and development firm The Peebles Corporation — the Golden State’s appeal is rapidly fading as residents pack up and head for the exits. “California, and especially Southern California, is the most difficult place to do business in the United States,” he stated bluntly in a Fox Business interview. “We were trying to build a $1.6 billion development in downtown LA, and stuck with it during the COVID crisis, and yet we could get no support from the government.” Don't miss Thanks to Jeff Bezos, you can now become a landlord for as little as $100 — and no, you don't have to deal with tenants or fix freezers. Here's how I'm 49 years old and have nothing saved for retirement — what should I do? Don't panic. Here are 5 of the easiest ways you can catch up (and fast) You don’t have to be a millionaire to gain access to this $1B private real estate fund. In fact, you can get started with as little as $10 — here’s how Peebles didn’t mince words, arguing that the state’s policies “were hurting businesses.” He also pointed to the growing wave of people leaving California. “People are fleeing, they have given up, and they're going to other places,” he said. “We're going to see more of an exodus out of California, because the quality of life has diminished as well.” The California exodus by the numbers Talk of a California exodus gained momentum during the pandemic, and although the pace has slowed, the outflow of residents hasn’t stopped. Where did they go? Texas topped the list. In 2023, 93,970 Californians relocated to the Lone Star State. In fact, Texas has consistently been the most popular destination for those leaving California: 107,546 Californians moved there in 2021 102,442 more followed in 2022 Arizona and Florida were also major draws, attracting 54,222 and 39,052 former Californians, respectively, during the most recent reporting period. Read more: Rich, young Americans are ditching the stormy stock market — here are the alternative assets they're banking on instead Rising costs of living — and how to hedge against them There are many theories about why so many Californians are leaving. High taxes are often cited — for example, neither Texas nor Florida imposes a state income tax. But perhaps just as important is the sky-high cost of living. Housing costs alone are enough to make headlines. According to data from real estate brokerage Redfin, the median home price in California currently stands at $859,700 — nearly twice the national median of $440,892. A recent Bankrate study found that a household in California needs an annual income of $213,447 to afford a typical home in the state. Yet real estate remains a popular investment choice for those looking to hedge against rising living costs. When inflation goes up, property values often climb as well, reflecting the higher costs of materials, labor and land. At the same time, rental income tends to rise, providing landlords with a revenue stream that adjusts with inflation. Over the past five years, the S&P CoreLogic Case-Shiller U.S. National Home Price NSA Index has surged more than 50%. These days, you don’t need to buy an entire property outright to benefit from real estate investing. Crowdfunding platforms like Arrived have made it easier than ever for everyday investors to gain exposure to this income-generating asset class. Backed by world class investors like Jeff Bezos, Arrived allows you to invest in shares of rental homes with as little as $100, all without the hassle of mowing lawns, fixing leaky faucets or handling difficult tenants. The process is simple: Browse a curated selection of homes that have been vetted for their appreciation and income potential. Once you find a property you like, select the number of shares you’d like to purchase, and then sit back as you start receiving positive rental income distributions from your investment. For accredited investors, Homeshares gives access to the $35 trillion U.S. home equity market, which has historically been the exclusive playground of institutional investors. With a minimum investment of $25,000, investors can gain direct exposure to hundreds of owner-occupied homes in top U.S. cities through their U.S. Home Equity Fund — without the headaches of buying, owning or managing property. With risk-adjusted target returns ranging from 14% to 17%, this approach provides an effective, hands-off way to invest in owner-occupied residential properties across regional markets. Another option is First National Realty Partners (FNRP), which allows accredited investors to diversify their portfolio through grocery-anchored commercial properties without taking on the responsibilities of being a landlord. With a minimum investment of $50,000, investors can own a share of properties leased by national brands like Whole Foods, Kroger and Walmart, which provide essential goods to their communities. Thanks to Triple Net (NNN) leases, accredited investors are able to invest in these properties without worrying about tenant costs cutting into their potential returns. Simply answer a few questions — including how much you would like to invest — to start browsing their full list of available properties. What to read next Here are the 6 levels of wealth for retirement-age Americans — are you near the top or bottom of the pyramid? This tiny hot Costco item has skyrocketed 74% in price in under 2 years — but now the retail giant is restricting purchases. Here’s how to buy the coveted asset in bulk Car insurance in America could climb to a stunning $2,502/year on average — but here’s how 2 minutes can save you more than $600 in 2025 Want an extra $1,300,000 when you retire? Dave Ramsey says this 7-step plan ‘works every single time’ to kill debt, get rich in America — and that ‘anyone’ can do it Money doesn’t have to be complicated — sign up for the free Moneywise newsletter for actionable finance tips and news you can use. Join now. This article provides information only and should not be construed as advice. It is provided without warranty of any kind. CoreWeave’s Cooling Stock Rally Faces Even Bigger Hurdles Ahead (Bloomberg) -- CoreWeave Inc.’s blistering rally has lost steam as the artificial intelligence infrastructure play faces a growing tangle of bearish catalysts that threaten to further unravel its meteoric rise. Shares of the Nvidia Corp.-backed company have tumbled 17% in less than a month after a nearly five-fold increase following its March initial public offering as investors soured on its ballooning valuation. Now it faces a wall of pressure with a series of share lockups expiring starting next month, allowing early shareholders to cash out. Meanwhile, its deal to buy Core Scientific Inc. for $9 billion drew mixed reviews, spurring at least three analysts downgrades. CoreWeave shares slipped 1% in early trading on Thursday. “We’re already giving CoreWeave a pass on so many things, not least of which is that it doesn’t make money and won’t for a while,” said Mark Malek, chief investment officer at Siebert Financial. “The lockup doesn’t change the story’s fundamentals, but it is another thing you’d have to give it a pass on if you’re buying now.” CoreWeave has become one of the most contentious stocks on Wall Street as its market value soared from less than $20 billion after its IPO to nearly $100 billion at its peak in June. The firm rents cloud-computing services to hyperscale-cloud operators like Microsoft Corp., making it a favorite of traders betting on a continuing wave of AI spending. But with the company not expected to turn a profit until the end of 2026, it has become a target of short sellers and value-minded investors who are betting it’s poised to fall back to earth. Even after its recent retreat, the stock trades at nearly 9 times sales estimated over the next 12 months, compared with roughly 5 times for the Nasdaq 100. CoreWeave did not respond to a Bloomberg News request for comment. Part of what has helped fuel CoreWeave’s rally is a relatively low number of shares available to trade, the result of the lockups and a small group of large shareholders, including Magnetar Financial and Coatue Management. Less than 13% of CoreWeave’s outstanding shares, about 47 million shares, are currently available to be traded. That will come to an end starting in August when an estimated 290 million shares will be freed up for trading. While many believe that the wave of shares hitting the market will exacerbate the current selloff in CoreWeave, some on Wall Street think it’s well positioned to weather the volatility as its second-quarter earnings approach. CoreWeave is expected to report a net loss of $236 million in the second quarter on revenue of $1.08 billion. “We think it will do significantly better than where CoreWeave is guiding,” said Mark Klein, chief executive officer at SuRo Capital. “You can’t just throw out traditional valuation metrics, but you have to appreciate how premium brands command premium valuations.” Still, much of Wall Street remains unconvinced. Only three of 22 analysts covering the stock recommend buying it, according to data compiled by Bloomberg. It’s average price target is more than 30% below where it closed on Wednesday, the worst expected return in a Bloomberg index of large capitalization stocks. Its all-stock deal to buy Core Scientific this week only served to further sour analysts on the stock. At least three analysts, including Needham, Mizuho and Stifel cut their ratings following the announcement. Both Needham and Mizuho cited valuation, while Stifel said it was positive on the deal longer term but called out overhangs from its pending dilution and lock-up expiry. “We see great potential, but right now all I see are mounting losses,” Siebert’s Malek said. “I’d rather leave money on the table than flush it down the toilet. Right now all I can do is watch the rally with a smile and say wow.”
Real estate CEO warns of growing ‘exodus’ as people have ‘given up’ on California — but where are they going? With its beautiful weather, breathtaking coastlines and vibrant culture, California has always held a special allure. But according to Don Peebles — founder, chairman and CEO of real estate investment and development firm The Peebles Corporation — the Golden State’s appeal is rapidly fading as residents pack up and head for the exits. “California, and especially Southern California, is the most difficult place to do business in the United States,” he stated bluntly in a Fox Business interview. “We were trying to build a $1.6 billion development in downtown LA, and stuck with it during the COVID crisis, and yet we could get no support from the government.” Don't miss Thanks to Jeff Bezos, you can now become a landlord for as little as $100 — and no, you don't have to deal with tenants or fix freezers. Here's how I'm 49 years old and have nothing saved for retirement — what should I do? Don't panic. Here are 5 of the easiest ways you can catch up (and fast) You don’t have to be a millionaire to gain access to this $1B private real estate fund. In fact, you can get started with as little as $10 — here’s how Peebles didn’t mince words, arguing that the state’s policies “were hurting businesses.” He also pointed to the growing wave of people leaving California. “People are fleeing, they have given up, and they're going to other places,” he said. “We're going to see more of an exodus out of California, because the quality of life has diminished as well.” The California exodus by the numbers Talk of a California exodus gained momentum during the pandemic, and although the pace has slowed, the outflow of residents hasn’t stopped. Where did they go? Texas topped the list. In 2023, 93,970 Californians relocated to the Lone Star State. In fact, Texas has consistently been the most popular destination for those leaving California: 107,546 Californians moved there in 2021 102,442 more followed in 2022 Arizona and Florida were also major draws, attracting 54,222 and 39,052 former Californians, respectively, during the most recent reporting period. Read more: Rich, young Americans are ditching the stormy stock market — here are the alternative assets they're banking on instead Rising costs of living — and how to hedge against them There are many theories about why so many Californians are leaving. High taxes are often cited — for example, neither Texas nor Florida imposes a state income tax. But perhaps just as important is the sky-high cost of living. Housing costs alone are enough to make headlines. According to data from real estate brokerage Redfin, the median home price in California currently stands at $859,700 — nearly twice the national median of $440,892. A recent Bankrate study found that a household in California needs an annual income of $213,447 to afford a typical home in the state. Yet real estate remains a popular investment choice for those looking to hedge against rising living costs. When inflation goes up, property values often climb as well, reflecting the higher costs of materials, labor and land. At the same time, rental income tends to rise, providing landlords with a revenue stream that adjusts with inflation. Over the past five years, the S&P CoreLogic Case-Shiller U.S. National Home Price NSA Index has surged more than 50%. These days, you don’t need to buy an entire property outright to benefit from real estate investing. Crowdfunding platforms like Arrived have made it easier than ever for everyday investors to gain exposure to this income-generating asset class. Backed by world class investors like Jeff Bezos, Arrived allows you to invest in shares of rental homes with as little as $100, all without the hassle of mowing lawns, fixing leaky faucets or handling difficult tenants. The process is simple: Browse a curated selection of homes that have been vetted for their appreciation and income potential. Once you find a property you like, select the number of shares you’d like to purchase, and then sit back as you start receiving positive rental income distributions from your investment. For accredited investors, Homeshares gives access to the $35 trillion U.S. home equity market, which has historically been the exclusive playground of institutional investors. With a minimum investment of $25,000, investors can gain direct exposure to hundreds of owner-occupied homes in top U.S. cities through their U.S. Home Equity Fund — without the headaches of buying, owning or managing property. With risk-adjusted target returns ranging from 14% to 17%, this approach provides an effective, hands-off way to invest in owner-occupied residential properties across regional markets. Another option is First National Realty Partners (FNRP), which allows accredited investors to diversify their portfolio through grocery-anchored commercial properties without taking on the responsibilities of being a landlord. With a minimum investment of $50,000, investors can own a share of properties leased by national brands like Whole Foods, Kroger and Walmart, which provide essential goods to their communities. Thanks to Triple Net (NNN) leases, accredited investors are able to invest in these properties without worrying about tenant costs cutting into their potential returns. Simply answer a few questions — including how much you would like to invest — to start browsing their full list of available properties. What to read next Here are the 6 levels of wealth for retirement-age Americans — are you near the top or bottom of the pyramid? This tiny hot Costco item has skyrocketed 74% in price in under 2 years — but now the retail giant is restricting purchases. Here’s how to buy the coveted asset in bulk Car insurance in America could climb to a stunning $2,502/year on average — but here’s how 2 minutes can save you more than $600 in 2025 Want an extra $1,300,000 when you retire? Dave Ramsey says this 7-step plan ‘works every single time’ to kill debt, get rich in America — and that ‘anyone’ can do it Money doesn’t have to be complicated — sign up for the free Moneywise newsletter for actionable finance tips and news you can use. Join now. This article provides information only and should not be construed as advice. It is provided without warranty of any kind. CoreWeave’s Cooling Stock Rally Faces Even Bigger Hurdles Ahead (Bloomberg) -- CoreWeave Inc.’s blistering rally has lost steam as the artificial intelligence infrastructure play faces a growing tangle of bearish catalysts that threaten to further unravel its meteoric rise. Shares of the Nvidia Corp.-backed company have tumbled 17% in less than a month after a nearly five-fold increase following its March initial public offering as investors soured on its ballooning valuation. Now it faces a wall of pressure with a series of share lockups expiring starting next month, allowing early shareholders to cash out. Meanwhile, its deal to buy Core Scientific Inc. for $9 billion drew mixed reviews, spurring at least three analysts downgrades. CoreWeave shares slipped 1% in early trading on Thursday. “We’re already giving CoreWeave a pass on so many things, not least of which is that it doesn’t make money and won’t for a while,” said Mark Malek, chief investment officer at Siebert Financial. “The lockup doesn’t change the story’s fundamentals, but it is another thing you’d have to give it a pass on if you’re buying now.” CoreWeave has become one of the most contentious stocks on Wall Street as its market value soared from less than $20 billion after its IPO to nearly $100 billion at its peak in June. The firm rents cloud-computing services to hyperscale-cloud operators like Microsoft Corp., making it a favorite of traders betting on a continuing wave of AI spending. But with the company not expected to turn a profit until the end of 2026, it has become a target of short sellers and value-minded investors who are betting it’s poised to fall back to earth. Even after its recent retreat, the stock trades at nearly 9 times sales estimated over the next 12 months, compared with roughly 5 times for the Nasdaq 100. CoreWeave did not respond to a Bloomberg News request for comment. Part of what has helped fuel CoreWeave’s rally is a relatively low number of shares available to trade, the result of the lockups and a small group of large shareholders, including Magnetar Financial and Coatue Management. Less than 13% of CoreWeave’s outstanding shares, about 47 million shares, are currently available to be traded. That will come to an end starting in August when an estimated 290 million shares will be freed up for trading. While many believe that the wave of shares hitting the market will exacerbate the current selloff in CoreWeave, some on Wall Street think it’s well positioned to weather the volatility as its second-quarter earnings approach. CoreWeave is expected to report a net loss of $236 million in the second quarter on revenue of $1.08 billion. “We think it will do significantly better than where CoreWeave is guiding,” said Mark Klein, chief executive officer at SuRo Capital. “You can’t just throw out traditional valuation metrics, but you have to appreciate how premium brands command premium valuations.” Still, much of Wall Street remains unconvinced. Only three of 22 analysts covering the stock recommend buying it, according to data compiled by Bloomberg. It’s average price target is more than 30% below where it closed on Wednesday, the worst expected return in a Bloomberg index of large capitalization stocks. Its all-stock deal to buy Core Scientific this week only served to further sour analysts on the stock. At least three analysts, including Needham, Mizuho and Stifel cut their ratings following the announcement. Both Needham and Mizuho cited valuation, while Stifel said it was positive on the deal longer term but called out overhangs from its pending dilution and lock-up expiry. “We see great potential, but right now all I see are mounting losses,” Siebert’s Malek said. “I’d rather leave money on the table than flush it down the toilet. Right now all I can do is watch the rally with a smile and say wow.”
考点 1:【exodus】 应译为 【人口外流 / 大规模迁出】 考点 2:【The Peebles Corporation】 应译为【 皮布尔斯公司】 考点 3:【Triple Net (NNN) leases】 应译为【 三重净租赁】 考点 4:【lockup expiry】 应译为 【锁定期到期】 考点 5:【sky-high cost of living 】应译为 【高得离谱的生活成本】 考点 6:【leave money on the table 】应译为 【错失潜在收益】 考点 7:【flush it down the toilet】 应译为 【打水漂 / 白白浪费】
2851e
垂类场景
房地产
89
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: When viewed as an elementary particle, the electron has spin and charge. When binding to the atomic nucleus, it also acquires an angular momentum quantum number corresponding to the quantized atomic orbital it occupies. Even if electrons in solids form bands and delocalize from the nuclei, in Mott insulators they retain their three fundamental quantum numbers: spin, charge and orbital1 . The hallmark of one-dimensional physics is a breaking up of the elementary electron into its separate degrees of freedom2 . The separation of the electron into independent quasi-particles that carry either spin (spinons) or charge (holons) was first observed fifteen years ago3 . Here we report observation of the separation of the orbital degree of freedom (orbiton) using resonant inelastic X-ray scattering on the one-dimensional Mott insulator Sr2CuO3. We resolve an orbiton separating itself from spinons and propagating through the lattice as a distinct quasiparticle with a substantial dispersion in energy over momentum, of about 0.2 electronvolts, over nearly one Brillouin zone. It was pointed out in the 1970s that in a solid not only the charge and spin of electrons can become ordered—leading to magnetism—but also the electrons’ orbital degree of freedom1 . This observation sparked a field that has gone on to produce a number of important results. Although a physical electron combines spin, charge and orbital, theoretically an electron can be considered a bound state of the three independent, fundamental quasi-particles: a spinon, carrying the electron’s spin; a holon (or chargon), carrying its charge; and an orbiton, carrying its orbital degree of freedom. A remarkable and fundamental property of one-dimensional (1D) systems is that electronic excitations break up into deconfined spinons and holons. This was predicted decades ago (ref. 2 and references therein) and confirmed in the mid 1990s by angle-resolved photoemission spectroscopy experiments3–5. The spin–charge separation is an example of particle fractionalization, a phenomenon in which the quantum numbers of quasi-particles are not multiples of those of the elementary particle, but fractions. This effect is one of the most unusual manifestations of collective quantum physics of interacting particles and is a profound concept that has found its way into a number of theories, for example that describing high-temperature superconductivity in copper oxides6,7. To search for the further fractionalization of the electron, we consider the excitation of a copper orbital degree of freedom in the antiferromagnetic spin-chain compound Sr2CuO3. The spin–orbital separation process that we are looking for is analogous to the spin–charge separation mechanism (Fig. 1b). The latter occurs, for instance, when an electron is annihilated, removing a single spin and leaving behind a hole in the antiferromagnetic chain. This hole can start to propagate freely only after exciting one spinon (a domain wall in the antiferromagnetic chain). Subsequently, the spinon can delocalize and separate itself completely from the holon. When instead of creating a hole, as typically is done in a photoemission experiment, an electron is excited from one copper 3d orbital to another, the phenomenon of spin–orbital separation can in principle occur (Fig. 1a). The orbiton created in this manner may also deconfine after exciting a spinon, thus splitting the electron into its orbital and spin degrees of freedom8 . Here we use high-resolution resonant inelastic X-ray scattering (RIXS) to search experimentally for spin–orbital separation in the quasi-1D copper oxide Sr2CuO3 (for material details, see Supplementary Information, section 1). We observe deconfinement of the spinon and orbiton during orbital excitation from the ground-state copper 3d orbital to an excited copper 3d xy or xz orbital (Fig. 1c–e). For simplicity, we will from henceforth use the so-called ‘hole’ language: although nominally there are nine electrons in the 3d orbitals of the Cu21 ion in Sr2CuO3, by using the ‘electron–hole’ transformation we can map the problem onto an effective system with one particle occupying a single 3d orbital (Supplementary Information, section 2). We measure an orbiton dispersion that is almost as large as the dispersion of the two-spinon continuum at low energies. As for spin–charge separation3–5, the orbiton dispersion has periodicity p (see Fig. 1c and see discussion below), which indicates the presence of an orbiton liberated from the spinon. We measured the orbital excitations of Sr2CuO3 using RIXS at the L3 edge of the copper ion. RIXS is a second-order scattering technique and can excite transitions between the copper 3d states of different symmetry (orbital excitations), owing to the involvement of two subsequent electric dipole transitions9,10 (Supplementary Information, section 3). With the unique capability of RIXS also to probe spin excitations11–13 and to vary the photon momentum transfer, the dispersion of orbital and spin excitations can be mapped out across the first Brillouin zone11–16. The experiments were carried out at the ADRESS beamline of the Swiss Light Source at the Paul Scherrer Institut17,18. For fixed momentum transfer, q, along the chains, peaks in the RIXS spectrum at constant energy transfer reveal the presence of charge-neutral elementary excitations and are visible in the RIXS intensity map of Sr2CuO3 across the copper L3 edge in Fig. 2a. The spectrum for which the incident energy was precisely tuned to the resonance maximum of the absorption spectrum is shown in Fig. 2b. In both plots, the excitations of the spin, orbital and charge degrees of freedom are indicated. The momentum dependence and, in particular, the dispersion of the spin and orbital excitations (Fig. 1c and Supplementary Fig. 2a) are indicative of their collective nature. For energy transfers of up to ,0.8 eV purely magnetic excitations are present, but the spectrum between ,1.5 and ,3.5 eV corresponds to excitations from the copper ground state to orbitals of symmetry (Fig. 1d, e). These peaks correspond toorbital excitations (called also d–d excitations), and not, for example, to charge transfer excitations, the intensity of which is non-zero for energies up to ,6 eV but here for the case of L-edge RIXS is at least an order of magnitude lower10. The orbital assignment of these excitations was unambiguously verified by comparing their energy with ab initio quantum chemistry cluster calculations19 (Supplementary Information, section 5, for detailed results). Zooming into the ‘magnetic’ part of Fig. 1c, between 0 and 0.8 eV in energy transfer, reveals strongly dispersing spin excitations. At the lower boundary the dispersion has period p, and at the continuum (upper boundary) it has period 2p (Fig. 3a). These RIXS data agree very well with recent inelastic neutron scattering studies on Sr2CuO3 (ref. 20). The simultaneous presence in the spectrum of a lower edge with period p and an upper one with period 2p indicates directly that in the spin chain the magnetic excitations with spin 1 break up and fractionalize into two-spinon (and higher-order) excitations that make up a continuum20, with each spinon having spin 1/2. These spectra confirm that RIXS for magnetic excitations probes the well-known spin dynamical structure factor as theoretically predicted11–13, in agreement with recent studies on TiOCl (ref. 21). The excellent statistics ofthe data further allow for a direct comparison of the RIXS line shapes with the exact two- and four-spinon dynamical structure factor of the spin-1/2 Heisenberg chain. In Fig. 3b, we show the fits for three selected momentum transfer values using the exact two- and four-spinon dynamical structure factor, S(q, v) (where q and v stand for the momentum and energy of the created excitation, respectively) in the representation of ref. 22. The obtained exchange coupling is in very good agreement with the value obtained from inelastic neutron scattering data20. Having unambiguously identified the fractionalized spinon excitations in the low-energy sector, we now concentrate on the orbital excitations spectrum in Fig. 4. We find that these are strongly momentum dependent and have a novel, distinct dispersion. This proves that the orbital excitations observed here are of collective nature. The xz excitation has the largest dispersion, of ,0.2 eV, and has a spectrum containing two peculiar components: a lower branch dispersing with periodicity p and, above that, an incoherent spectrum with a double-oval shape. This spectrum is strikingly similar to seemingly unrelated angle-resolved photoemission spectra of 1D copper oxides, which evidence spin–charge separation (see, for example, fig. 3 of ref. 3). This is an indication that the observed orbital dispersion is related to an analogous separation of degrees of freedom. To test this conjecture, we derived a microscopic model that describes the spin–orbital interactions in Sr2CuO3 (Methods). The low-energy Kugel’–Khomskii Hamiltonian for this was obtained8,23 from the charge transfer model of Sr2CuO3 in ref. 24. The crucial part of the Hamiltonian, responsible for the xz orbital propagation, where J and JO are respectively the spin and orbital exchange constants, both of which are fixed by the charge transfer model24; EO is the xz orbital on-site energy; c z js and cjs are operators that respectively annihilate and create particles in xz orbital with spin s; Sj is the spin of the particle in x2 2 y 2 orbital; 1 2 nj counts the number of particles in xz orbital; and ‘h.c.’ denotes Hermitian conjugate. The first term in the Hamiltonian describes the propagation of orbital excitations through the lattice. The second term represents the usual Heisenberg interaction between spins, which vanishes on bonds where an orbital excitation is present (Methods To quantify the spectral weights within the orbiton–spinon continuum, we have calculated for the JO–J Hamiltonian the orbital excitation Green’s function by exactly diagonalizing the Hamiltonian on 28 lattice sites (as in the spin–charge separation studies3 , finite-size effects are negligible (they are estimated at ,0.01 eV)). From this, we calculate the RIXS spectrum following ref. 13, that is, by expressing the RIXS amplitude as a product of the single-ion local RIXS effective operator and the orbital excitation Green’s function (Methods). We find excellent agreement between theory and experiment (Fig. 4): both the sine-like xz orbiton dispersion and the xz spinon– orbiton continuum (the ‘double-oval’ incoherent spectrum), which are the hallmarks of spin–orbital separation, are present in the theory. The calculations also show that, in contrast to the xz orbital, the xy orbital has a small dispersion and that the excitations of both the orbital and the yz orbital are dispersionless, as is observed experimentally. This is an independent merit of the model because no fitting of dispersions to experimental data is involved. The large orbiton dispersion observed in this study is the key feature that distinguishes Sr2CuO3 from other systems with orbital excitations27–29, and relies on the 1D character of Sr2CuO3. In a system of higher dimensionality, orbitons interact with magnetic excitations, which tend to slow them down and thus reduce their dispersion. In one dimension, orbitons can avoid these renormalization effects by means of spin–orbital separation
When viewed as an elementary particle, the electron has spin and charge. When binding to the atomic nucleus, it also acquires an angular momentum quantum number corresponding to the quantized atomic orbital it occupies. Even if electrons in solids form bands and delocalize from the nuclei, in Mott insulators they retain their three fundamental quantum numbers: spin, charge and orbital1 . The hallmark of one-dimensional physics is a breaking up of the elementary electron into its separate degrees of freedom2 . The separation of the electron into independent quasi-particles that carry either spin (spinons) or charge (holons) was first observed fifteen years ago3 . Here we report observation of the separation of the orbital degree of freedom (orbiton) using resonant inelastic X-ray scattering on the one-dimensional Mott insulator Sr2CuO3. We resolve an orbiton separating itself from spinons and propagating through the lattice as a distinct quasiparticle with a substantial dispersion in energy over momentum, of about 0.2 electronvolts, over nearly one Brillouin zone. It was pointed out in the 1970s that in a solid not only the charge and spin of electrons can become ordered—leading to magnetism—but also the electrons’ orbital degree of freedom1 . This observation sparked a field that has gone on to produce a number of important results. Although a physical electron combines spin, charge and orbital, theoretically an electron can be considered a bound state of the three independent, fundamental quasi-particles: a spinon, carrying the electron’s spin; a holon (or chargon), carrying its charge; and an orbiton, carrying its orbital degree of freedom. A remarkable and fundamental property of one-dimensional (1D) systems is that electronic excitations break up into deconfined spinons and holons. This was predicted decades ago (ref. 2 and references therein) and confirmed in the mid 1990s by angle-resolved photoemission spectroscopy experiments3–5. The spin–charge separation is an example of particle fractionalization, a phenomenon in which the quantum numbers of quasi-particles are not multiples of those of the elementary particle, but fractions. This effect is one of the most unusual manifestations of collective quantum physics of interacting particles and is a profound concept that has found its way into a number of theories, for example that describing high-temperature superconductivity in copper oxides6,7. To search for the further fractionalization of the electron, we consider the excitation of a copper orbital degree of freedom in the antiferromagnetic spin-chain compound Sr2CuO3. The spin–orbital separation process that we are looking for is analogous to the spin–charge separation mechanism (Fig. 1b). The latter occurs, for instance, when an electron is annihilated, removing a single spin and leaving behind a hole in the antiferromagnetic chain. This hole can start to propagate freely only after exciting one spinon (a domain wall in the antiferromagnetic chain). Subsequently, the spinon can delocalize and separate itself completely from the holon. When instead of creating a hole, as typically is done in a photoemission experiment, an electron is excited from one copper 3d orbital to another, the phenomenon of spin–orbital separation can in principle occur (Fig. 1a). The orbiton created in this manner may also deconfine after exciting a spinon, thus splitting the electron into its orbital and spin degrees of freedom8 . Here we use high-resolution resonant inelastic X-ray scattering (RIXS) to search experimentally for spin–orbital separation in the quasi-1D copper oxide Sr2CuO3 (for material details, see Supplementary Information, section 1). We observe deconfinement of the spinon and orbiton during orbital excitation from the ground-state copper 3d orbital to an excited copper 3d xy or xz orbital (Fig. 1c–e). For simplicity, we will from henceforth use the so-called ‘hole’ language: although nominally there are nine electrons in the 3d orbitals of the Cu21 ion in Sr2CuO3, by using the ‘electron–hole’ transformation we can map the problem onto an effective system with one particle occupying a single 3d orbital (Supplementary Information, section 2). We measure an orbiton dispersion that is almost as large as the dispersion of the two-spinon continuum at low energies. As for spin–charge separation3–5, the orbiton dispersion has periodicity p (see Fig. 1c and see discussion below), which indicates the presence of an orbiton liberated from the spinon. We measured the orbital excitations of Sr2CuO3 using RIXS at the L3 edge of the copper ion. RIXS is a second-order scattering technique and can excite transitions between the copper 3d states of different symmetry (orbital excitations), owing to the involvement of two subsequent electric dipole transitions9,10 (Supplementary Information, section 3). With the unique capability of RIXS also to probe spin excitations11–13 and to vary the photon momentum transfer, the dispersion of orbital and spin excitations can be mapped out across the first Brillouin zone11–16. The experiments were carried out at the ADRESS beamline of the Swiss Light Source at the Paul Scherrer Institut17,18. For fixed momentum transfer, q, along the chains, peaks in the RIXS spectrum at constant energy transfer reveal the presence of charge-neutral elementary excitations and are visible in the RIXS intensity map of Sr2CuO3 across the copper L3 edge in Fig. 2a. The spectrum for which the incident energy was precisely tuned to the resonance maximum of the absorption spectrum is shown in Fig. 2b. In both plots, the excitations of the spin, orbital and charge degrees of freedom are indicated. The momentum dependence and, in particular, the dispersion of the spin and orbital excitations (Fig. 1c and Supplementary Fig. 2a) are indicative of their collective nature. For energy transfers of up to ,0.8 eV purely magnetic excitations are present, but the spectrum between ,1.5 and ,3.5 eV corresponds to excitations from the copper ground state to orbitals of symmetry (Fig. 1d, e). These peaks correspond toorbital excitations (called also d–d excitations), and not, for example, to charge transfer excitations, the intensity of which is non-zero for energies up to ,6 eV but here for the case of L-edge RIXS is at least an order of magnitude lower10. The orbital assignment of these excitations was unambiguously verified by comparing their energy with ab initio quantum chemistry cluster calculations19 (Supplementary Information, section 5, for detailed results). Zooming into the ‘magnetic’ part of Fig. 1c, between 0 and 0.8 eV in energy transfer, reveals strongly dispersing spin excitations. At the lower boundary the dispersion has period p, and at the continuum (upper boundary) it has period 2p (Fig. 3a). These RIXS data agree very well with recent inelastic neutron scattering studies on Sr2CuO3 (ref. 20). The simultaneous presence in the spectrum of a lower edge with period p and an upper one with period 2p indicates directly that in the spin chain the magnetic excitations with spin 1 break up and fractionalize into two-spinon (and higher-order) excitations that make up a continuum20, with each spinon having spin 1/2. These spectra confirm that RIXS for magnetic excitations probes the well-known spin dynamical structure factor as theoretically predicted11–13, in agreement with recent studies on TiOCl (ref. 21). The excellent statistics ofthe data further allow for a direct comparison of the RIXS line shapes with the exact two- and four-spinon dynamical structure factor of the spin-1/2 Heisenberg chain. In Fig. 3b, we show the fits for three selected momentum transfer values using the exact two- and four-spinon dynamical structure factor, S(q, v) (where q and v stand for the momentum and energy of the created excitation, respectively) in the representation of ref. 22. The obtained exchange coupling is in very good agreement with the value obtained from inelastic neutron scattering data20. Having unambiguously identified the fractionalized spinon excitations in the low-energy sector, we now concentrate on the orbital excitations spectrum in Fig. 4. We find that these are strongly momentum dependent and have a novel, distinct dispersion. This proves that the orbital excitations observed here are of collective nature. The xz excitation has the largest dispersion, of ,0.2 eV, and has a spectrum containing two peculiar components: a lower branch dispersing with periodicity p and, above that, an incoherent spectrum with a double-oval shape. This spectrum is strikingly similar to seemingly unrelated angle-resolved photoemission spectra of 1D copper oxides, which evidence spin–charge separation (see, for example, fig. 3 of ref. 3). This is an indication that the observed orbital dispersion is related to an analogous separation of degrees of freedom. To test this conjecture, we derived a microscopic model that describes the spin–orbital interactions in Sr2CuO3 (Methods). The low-energy Kugel’–Khomskii Hamiltonian for this was obtained8,23 from the charge transfer model of Sr2CuO3 in ref. 24. The crucial part of the Hamiltonian, responsible for the xz orbital propagation, where J and JO are respectively the spin and orbital exchange constants, both of which are fixed by the charge transfer model24; EO is the xz orbital on-site energy; c z js and cjs are operators that respectively annihilate and create particles in xz orbital with spin s; Sj is the spin of the particle in x2 2 y 2 orbital; 1 2 nj counts the number of particles in xz orbital; and ‘h.c.’ denotes Hermitian conjugate. The first term in the Hamiltonian describes the propagation of orbital excitations through the lattice. The second term represents the usual Heisenberg interaction between spins, which vanishes on bonds where an orbital excitation is present (Methods To quantify the spectral weights within the orbiton–spinon continuum, we have calculated for the JO–J Hamiltonian the orbital excitation Green’s function by exactly diagonalizing the Hamiltonian on 28 lattice sites (as in the spin–charge separation studies3 , finite-size effects are negligible (they are estimated at ,0.01 eV)). From this, we calculate the RIXS spectrum following ref. 13, that is, by expressing the RIXS amplitude as a product of the single-ion local RIXS effective operator and the orbital excitation Green’s function (Methods). We find excellent agreement between theory and experiment (Fig. 4): both the sine-like xz orbiton dispersion and the xz spinon– orbiton continuum (the ‘double-oval’ incoherent spectrum), which are the hallmarks of spin–orbital separation, are present in the theory. The calculations also show that, in contrast to the xz orbital, the xy orbital has a small dispersion and that the excitations of both the orbital and the yz orbital are dispersionless, as is observed experimentally. This is an independent merit of the model because no fitting of dispersions to experimental data is involved. The large orbiton dispersion observed in this study is the key feature that distinguishes Sr2CuO3 from other systems with orbital excitations27–29, and relies on the 1D character of Sr2CuO3. In a system of higher dimensionality, orbitons interact with magnetic excitations, which tend to slow them down and thus reduce their dispersion. In one dimension, orbitons can avoid these renormalization effects by means of spin–orbital separation
考点1. "two-spinon continuum"推荐翻译成"双自旋子连续谱"。 考点2. "Brillouin zone"推荐翻译成"布里渊区",不可译为“布里墟区”。 考点3:“an orbiton liberated from the spinon”中liberated推荐译为“脱离”、“分离”或“摆脱”,不可译为“解放”
29f57
学术论文
自然科学
187
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 世界上每四瓶番茄酱中,就有一瓶产自新疆,这既是新疆的贡献,也是新疆的自豪。 新疆番茄为什么“红”? 地理条件得天独厚 被誉为番茄的“黄金产地” 从异域传入中国的番茄非常喜欢充足的日照和剧烈的温差,而这些,新疆统统满足。 日照丰富是新疆黄金产区的一大特色,新疆全年日照时数可达2550~3300小时,在全国各省区中位居第一。天山雪水的灌溉,成为番茄品质的加分项,因为天山雪水富含多种矿物元素,为生长在这里的番茄赋予更多矿物质和微量元素。 新疆优越的地理条件,成就了品质绝佳的番茄,味道与营养兼具,成为越来越多人的首选。也正因此,新疆成为全球三大番茄核心产区之一,而中国也是世界第一大番茄制品出口国。 育种创新 让新疆成为番茄遗传资源富集区 “全世界一共研究出13个番茄种的基因组,我们现在掌握11个。”新疆农业科学院副院长、研究员、加工番茄生物育种创新团队学科带头人余庆辉说。据介绍,该研究利用第三代单分子实时测序技术、全基因组光学图谱技术和高通量染色体构象捕获测序技术对11个番茄种进行基因测序组装,组装出11个染色体水平高质量基因组,打开了番茄属遗传资源领域研究的宝库,使新疆从种植大区一跃成为世界番茄遗传资源富集区。 近年来,借助多元化的育种技术,育种专家对野生番茄持续进行改良,培育出了各具特色的栽培番茄品种,使番茄的抗病性、产量、固形物含量等特性显著放大,满足了产业发展需求。“有时产量提升了,却失去了风味,很难有兼顾各方需求的完美品种。”新疆农科院加工番茄生物育种创新团队副研究员李宁说,很多人反映现在的番茄没有过去好吃了,其实就是在品种选育的过程中,为了突出产量等特性,人为导致丢失了风味特性的一种结果。 自2018年研究工作启动以来,这个团队收集了8个野生番茄种、1个番茄近源野生种和2个栽培番茄代表性品种,构建了番茄属泛基因组,解析了番茄属基因组特征,重构了番茄属系统发生关系,为下一步基因的分型研究提供了平台。 “三高两少”优质好吃 百搭做法融出美妙滋味 由于自然条件的优势,新疆番茄具有“三高两少”的品质优势,“三高”指红色素含量高、可溶性固形物含量高、单产高;“两少”是指病虫害少、霉菌少。这就大大减少裂果、霉烂果的发生,提高整体的番茄品质。番茄霉菌视野小于25%,最低可在12%以下,远远低于我国和国外一些国家的规定标准(加拿大50%、中国40%)。于是番茄酱“两高一大”的特点就凸显出来了,“两高”是指新鲜度指标和粘度指标在世界市场上最高,“一大”是指番茄酱的理想化指标,即颗粒度大可以分装制造出最优质的沙司。 新疆生产建设兵团第二师21团职工在戈壁滩晒场晾晒番茄干。今年,该团晾晒番茄干100余吨,主要销往东南亚、非洲等地。 新疆番茄酱粘度好、酸甜可口、富有鲜味,其独有的酸甜口感在番茄火锅、番茄牛腩、番茄鸡翅、番茄大虾等特色菜品中起到了至关重要的作用,深受消费者的喜爱。 在新疆,各类地道美食自然也少不了番茄的装点。劲道的新疆拌面包裹上一层酸爽的番茄酱,瞬间惊艳味蕾;在红透了的新疆炒米粉里,有了番茄才辣而不燥。番茄就像一个若隐若现的“百变厨神”,放在哪里,哪里就熠熠生辉。
世界上每四瓶番茄酱中,就有一瓶产自新疆,这既是新疆的贡献,也是新疆的自豪。 新疆番茄为什么“红”? 地理条件得天独厚 被誉为番茄的“黄金产地” 从异域传入中国的番茄非常喜欢充足的日照和剧烈的温差,而这些,新疆统统满足。 日照丰富是新疆黄金产区的一大特色,新疆全年日照时数可达2550~3300小时,在全国各省区中位居第一。天山雪水的灌溉,成为番茄品质的加分项,因为天山雪水富含多种矿物元素,为生长在这里的番茄赋予更多矿物质和微量元素。 新疆优越的地理条件,成就了品质绝佳的番茄,味道与营养兼具,成为越来越多人的首选。也正因此,新疆成为全球三大番茄核心产区之一,而中国也是世界第一大番茄制品出口国。 育种创新 让新疆成为番茄遗传资源富集区 “全世界一共研究出13个番茄种的基因组,我们现在掌握11个。”新疆农业科学院副院长、研究员、加工番茄生物育种创新团队学科带头人余庆辉说。据介绍,该研究利用第三代单分子实时测序技术、全基因组光学图谱技术和高通量染色体构象捕获测序技术对11个番茄种进行基因测序组装,组装出11个染色体水平高质量基因组,打开了番茄属遗传资源领域研究的宝库,使新疆从种植大区一跃成为世界番茄遗传资源富集区。 近年来,借助多元化的育种技术,育种专家对野生番茄持续进行改良,培育出了各具特色的栽培番茄品种,使番茄的抗病性、产量、固形物含量等特性显著放大,满足了产业发展需求。“有时产量提升了,却失去了风味,很难有兼顾各方需求的完美品种。”新疆农科院加工番茄生物育种创新团队副研究员李宁说,很多人反映现在的番茄没有过去好吃了,其实就是在品种选育的过程中,为了突出产量等特性,人为导致丢失了风味特性的一种结果。 自2018年研究工作启动以来,这个团队收集了8个野生番茄种、1个番茄近源野生种和2个栽培番茄代表性品种,构建了番茄属泛基因组,解析了番茄属基因组特征,重构了番茄属系统发生关系,为下一步基因的分型研究提供了平台。 “三高两少”优质好吃 百搭做法融出美妙滋味 由于自然条件的优势,新疆番茄具有“三高两少”的品质优势,“三高”指红色素含量高、可溶性固形物含量高、单产高;“两少”是指病虫害少、霉菌少。这就大大减少裂果、霉烂果的发生,提高整体的番茄品质。番茄霉菌视野小于25%,最低可在12%以下,远远低于我国和国外一些国家的规定标准(加拿大50%、中国40%)。于是番茄酱“两高一大”的特点就凸显出来了,“两高”是指新鲜度指标和粘度指标在世界市场上最高,“一大”是指番茄酱的理想化指标,即颗粒度大可以分装制造出最优质的沙司。 新疆生产建设兵团第二师21团职工在戈壁滩晒场晾晒番茄干。今年,该团晾晒番茄干100余吨,主要销往东南亚、非洲等地。 新疆番茄酱粘度好、酸甜可口、富有鲜味,其独有的酸甜口感在番茄火锅、番茄牛腩、番茄鸡翅、番茄大虾等特色菜品中起到了至关重要的作用,深受消费者的喜爱。 在新疆,各类地道美食自然也少不了番茄的装点。劲道的新疆拌面包裹上一层酸爽的番茄酱,瞬间惊艳味蕾;在红透了的新疆炒米粉里,有了番茄才辣而不燥。番茄就像一个若隐若现的“百变厨神”,放在哪里,哪里就熠熠生辉。
考点1:“新疆番茄为什么“红””中,红不可直译为red,结合语境推荐译为hot。 考点2:黄金产地应该译为“a prime growing area”,此处采用比喻手法。 考点3:泛基因组应译为“pangenome”,是基因组学术语。 考点4:新疆生产建设兵团第二师21团职工应译为“employees from the 21st Regiment of the 2nd Division of the Xinjiang Production and Construction Corps (XPCC)“。 考点5:百变厨神推荐译为“a versatile culinary master”。
2a470
新闻资讯
新闻报道
29
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 调查过程中常出现数据缺失,在这种情形下不能直接将含缺失值的个体信息全部删除而只利用数据完全的样本数据,因为数据缺失与否往往与该数据的具体值系统地相关,比如关心的因变量是受访者的年收入,高收入群体可能不愿意将自己的收入告知他人,这种情况造成的数据缺失就不能采用直接删除法,否则会造成估计结果偏低。为此收集尽量多的协变量数据就很有意义。可以采用插补法(包括回归插补法、近邻插补法)或似然法(通过建立因变量、协变量和缺失指示变量的分布模型)[1]。在这些方法中,以往研究都假设协变量数据没有缺失,回归插补法通过有回答的样本数据建立模型,代入因变量数据缺失样本的协变量数据以估计因变量值。近邻插补法通过定义距离函数,考虑用与因变量缺失个体的协变量最相似有数据的个体因变量值插补。似然法则引入数据缺失机制,通过建模得到因变量与缺失指示变量的联合分布,进而得到参数的极大似然估计。实际调查中,协变量缺失的情况也是非常常见的,例如出于隐私考虑不愿公布年龄或婚姻状况,由于知识水平所限无法回答某些超出其知识范围的问题,也可能某些协变量收集难度大成本高,调查人员选择只收集一部分受访者的信息。协变量缺失时因变量可能没有缺失,也可能协变量和因变量同时缺失,这就大大增加了协变量关于因变量回归系数估计的难度。但目前国内外关于协变量缺失情形下尤其是缺失机制是非随机缺失时参数的估计方法研究还不多,Little(1992)[2]研究了回归分析中缺失协变量的处理方法,但其研究局限于因变量与协变量之间满足线性回归关系,且假设因变量Y与协变量向量X的联合分布。为多元正态分布,Horton和Laird(1999)[3]研究了缺失机制为MAR的属性协变量缺失问题,通过建立广义线性模型,用极大似然法进行参数估计,Michiels等(1999)[4]用选择模型和模式混合模型建立似然函数,用极大似然法估计属性协变量数据缺失时的总体参数。在软件功能日益强大的今天,数据扩充算法、EM算法、Gibbs抽样法等迭代算法应用日益广泛。SAS Proc MI可以采用多重插补法进行协变量数据缺失情形下的参数估计。R有多个软件包可以进行缺失数据统计分析,ACD包可以在因变量数据缺失时进行属性数据分析,mvnmle 包在因变量和协变量联合分布为多元正态分布时,进行协变量数据缺失情形下的参数极大似然估计,MICE包是R中目前最常用的用于缺失数据分析的软件包,可以进行多变量缺失数据的多重插补,在多个协变量都可能存在缺失值时,使用MICE包中的mice函数,通过变量之间的关系预测缺失数据,利用蒙特卡洛方法生成多个完整数据集存在imp中,再对imp进行线性回归,最后用pool函数对回归结果进行汇总。本文尝试引入以上统计计算方法,采用多重插补法、Bayes法和极大似然法,在缺失机制为MAR或NMAR情形下,估计调查数据中协变量数据缺失情形下,协变量关于因变量影响大小的回归系数。 2.1 完全数据法 该方法将所有含缺失值的受访者删除,仅保留完全数据样本进行统计分析,这种方法仅当缺失机制为MCAR时可以得到参数无偏估计,缺失机制为MAR或NMAR时得到的估计量有偏,若协变量数目较多,即使某受访者只有一个协变量数据缺失,仍需要将其所有数据删去,这样可能导致可用的数据量不足,造成大量的信息浪费。虽然这种方法经常采用,但并不值得推荐。本文尝试采用EM算法、Gibbs抽样法和数据扩充算法等统计计算方法,采用多重插补法、Bayes 法和极大似然法,在缺失机制为MAR或NMAR情形下,估计调查数据中协变量数据缺失情形下,反映协变量X对因变量y 影响大小的回归系数。研究可以进一步推广到协变量和因变量同时缺失的情形下定义缺失机制、建立似然模型并采用统计算法进行回归系数估计。还可以推广到对多个体多时点纵向调查,在协变量缺失的情形下,引入统计算法,进行回归系数估计。
调查过程中常出现数据缺失,在这种情形下不能直接将含缺失值的个体信息全部删除而只利用数据完全的样本数据,因为数据缺失与否往往与该数据的具体值系统地相关,比如关心的因变量是受访者的年收入,高收入群体可能不愿意将自己的收入告知他人,这种情况造成的数据缺失就不能采用直接删除法,否则会造成估计结果偏低。为此收集尽量多的协变量数据就很有意义。可以采用插补法(包括回归插补法、近邻插补法)或似然法(通过建立因变量、协变量和缺失指示变量的分布模型)[1]。在这些方法中,以往研究都假设协变量数据没有缺失,回归插补法通过有回答的样本数据建立模型,代入因变量数据缺失样本的协变量数据以估计因变量值。近邻插补法通过定义距离函数,考虑用与因变量缺失个体的协变量最相似有数据的个体因变量值插补。似然法则引入数据缺失机制,通过建模得到因变量与缺失指示变量的联合分布,进而得到参数的极大似然估计。实际调查中,协变量缺失的情况也是非常常见的,例如出于隐私考虑不愿公布年龄或婚姻状况,由于知识水平所限无法回答某些超出其知识范围的问题,也可能某些协变量收集难度大成本高,调查人员选择只收集一部分受访者的信息。协变量缺失时因变量可能没有缺失,也可能协变量和因变量同时缺失,这就大大增加了协变量关于因变量回归系数估计的难度。但目前国内外关于协变量缺失情形下尤其是缺失机制是非随机缺失时参数的估计方法研究还不多,Little(1992)[2]研究了回归分析中缺失协变量的处理方法,但其研究局限于因变量与协变量之间满足线性回归关系,且假设因变量Y与协变量向量X的联合分布。为多元正态分布,Horton和Laird(1999)[3]研究了缺失机制为MAR的属性协变量缺失问题,通过建立广义线性模型,用极大似然法进行参数估计,Michiels等(1999)[4]用选择模型和模式混合模型建立似然函数,用极大似然法估计属性协变量数据缺失时的总体参数。在软件功能日益强大的今天,数据扩充算法、EM算法、Gibbs抽样法等迭代算法应用日益广泛。SAS Proc MI可以采用多重插补法进行协变量数据缺失情形下的参数估计。R有多个软件包可以进行缺失数据统计分析,ACD包可以在因变量数据缺失时进行属性数据分析,mvnmle 包在因变量和协变量联合分布为多元正态分布时,进行协变量数据缺失情形下的参数极大似然估计,MICE包是R中目前最常用的用于缺失数据分析的软件包,可以进行多变量缺失数据的多重插补,在多个协变量都可能存在缺失值时,使用MICE包中的mice函数,通过变量之间的关系预测缺失数据,利用蒙特卡洛方法生成多个完整数据集存在imp中,再对imp进行线性回归,最后用pool函数对回归结果进行汇总。本文尝试引入以上统计计算方法,采用多重插补法、Bayes法和极大似然法,在缺失机制为MAR或NMAR情形下,估计调查数据中协变量数据缺失情形下,协变量关于因变量影响大小的回归系数。 2.1 完全数据法 该方法将所有含缺失值的受访者删除,仅保留完全数据样本进行统计分析,这种方法仅当缺失机制为MCAR时可以得到参数无偏估计,缺失机制为MAR或NMAR时得到的估计量有偏,若协变量数目较多,即使某受访者只有一个协变量数据缺失,仍需要将其所有数据删去,这样可能导致可用的数据量不足,造成大量的信息浪费。虽然这种方法经常采用,但并不值得推荐。本文尝试采用EM算法、Gibbs抽样法和数据扩充算法等统计计算方法,采用多重插补法、Bayes 法和极大似然法,在缺失机制为MAR或NMAR情形下,估计调查数据中协变量数据缺失情形下,反映协变量X对因变量y 影响大小的回归系数。研究可以进一步推广到协变量和因变量同时缺失的情形下定义缺失机制、建立似然模型并采用统计算法进行回归系数估计。还可以推广到对多个体多时点纵向调查,在协变量缺失的情形下,引入统计算法,进行回归系数估计。
考点1:"直接删除法"必须译为" Complete case analysis" 考点2:“大量的信息浪费”不可直译为“a lot of information waste”,推荐译为“a significant loss of information” 考点3:“估计结果偏低”中的“偏低”不可直译为“be low”,可以译为“be biased downward”或“be artificially low”
2b9b9
学术论文
自然科学
189
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: In contemporary artificial intelligence research, the problem of layout (Layout) generation—the automated arrangement of visually distinct objects or functional elements within a designed frame—remains of paramount practical and theoretical significance. Layouts underpin a broad spectrum of applications ranging from magazine publishing and poster production to the design of complex user interfaces in mobile and web environments. A well-constructed layout not only directs user attention and guides reading order, but also encodes semantic relationships among visual or textual components, incorporates constraints dictated by content attributes, and adheres to general and domain-specific aesthetic principles. Despite rapid progress in machine learning, the manual construction of graphic layouts remains an arduous and resource-intensive process. Designers are required to orchestrate element positioning, sizing, and alignment in accordance with both objective communicative needs and subjective artistic judgments, a task that can become intractable as the complexity and variability of both content and user-specified constraints increase. This motivates the development of intelligent systems for automatic layout generation (Automatic Layout Generation), allowing efficient synthesis of designs that satisfy prescribed structural and semantic requirements. Early computational methods for layout synthesis often relied on the formalization of expert rules as handcrafted energy functions or procedural grammars, encoding a priori notions of spatial harmony and alignment into optimization objectives. However, such rule-based or heuristic-driven systems lack scalability and generality, particularly when confronted with data distributions that exhibit significant heterogeneity or intricate relationships between design eleents (Design Elements). They are also limited by their dependence on manual encoding of preferences—a process difficult to generalize across diverse domains such as editorial layout, digital ads, and mobile UI screens. Recent advances in deep generative models have revolutionized automatic layout generation by allowing systems to learn directly from large-scale, annotated datasets . Two especially influential approaches in this context are Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) . GANs are formulated as two-player zero-sum games between a generator that synthesizes candidate samples and a discriminator that assesses the realism of those samples relative to observed data. Such adversarial training fosters the creation of highly plausible, often photorealistic outputs. In contrast, VAEs employ probabilistic encodings and Bayesian inference to model latent variable distributions over input data, enabling diversity and smooth interpolation between generated layouts. Each method, however, exhibits distinct limitations in the context of layout synthesis . GANs are infamous for issues such as training instability , mode collapse , and incomplete distribution coverage ; these can impede their ability to accurately recover the rich diversity of real-world layouts, especially in visually cluttered or structurally complex domains. VAEs , while often easier to train and able to capture a broader range of plausible alternatives, can suffer from trade-offs between sharpness and diversity , with generated layouts sometimes lacking the visual fidelity required for professional design. More recently, research attention has shifted to a new class of generative models—the Diffusion Models (Diffusion Models)—notably represented by the Denoising Diffusion Probabilistic Model (DDPM) and related architectures. Diffusion models pose data generation as a progressive stochastic process, starting from pure Gaussian noise and iteratively denoising to synthesize high-quality structured outputs. The forward process gradually injects noise into training data according to a predefined noise schedule, while the reverse process (reverse process) employs a neural network (typically a U-Net ) to remove noise in discrete steps, reconstructing data samples that closely approximate the target distribution. Key advantages of diffusion models include the ability to achieve a stationary training objective (stationary training objective)—circumventing adversarial oscillation—as well as the capacity to faithfully cover the underlying data distribution . Moreover, these models afford strong sample diversity and permit straightforward conditioning on auxiliary information, making them theoretically attractive for the complex, multi-constraint setting of graphical layout generation . However, several key challenges must be addressed to adapt diffusion models to the specific structure of layout data . First, unlike grid-structured images or time-sequenced signals, a graphic layout typically comprises a variable-length, unordered set of elements, each described by both discrete (e.g., class label) and continuous (e.g., size, coordinates) attributes. Conventional convolutional neural networks (CNNs) , which excel at learning regular pixel grids but presume spatial locality, are generally ill-suited for processing such sets, especially when explicit modeling of relationships and attribute constraints is required. Second, the semantics of a layout are not solely a function of its individual components but are emergent from the interdependencies among elements, meaning that conditional generation demands sophisticated mechanisms for relational reasoning . Addressing these obstacles, contemporary research has advocated for the use of Transformer-based architectures (Transformer-based architectures) as the neural backbone in layout diffusion models. The self-attention mechanism intrinsic to Transformer layers is adept at modeling dependencies in unordered and variable-sized input sets, enabling efficient aggregation of global context and complex attribute interactions. Specifically, in the context of layout diffusion , a Transformer-based denoiser can predict noise vectors for each element token based on global information , facilitating accurate and flexible reverse diffusion even when layout cardinality varies significantly across samples. In the conditional generation setting, such architectures further allow for the inclusion of rich attribute embeddings —capturing category labels, style codes, and textual descriptors—so that the generated layouts respect user-controlled constraints and design intents. Crucially, because the arrangement of layout elements is unordered by nature, these models typically omit positional encoding (positional encoding), diverging from mainstream practices in natural language processing or image modeling where sequence order is semantically essential. To formally describe the layout modeling process, suppose that a layout instance L is a set of N elements, with each element ei represented as a tuple of continuous geometry gi (position, width, height) and discrete or textual attributes fi (e.g., “heading”, “button”, or “price”). For consistency and stability, geometric parameters are normalized onto a fixed unit hypercube. The generative apparatus—namely, a conditional denoising diffusion probabilistic model (conditional DDPM)—defines a forward process, which stepwise corrupts g by adding Gaussian noise, and a reverse process, which sequentially denoises g back to its original configuration, all conditioned on f. The conditional layout denoiser (cLayoutDenoiser) is realized as a sequence of stacked Transformer blocks . For each sample and each reverse diffusion time step t, the model receives as input (i) geometry embeddings , i.e., learned projections of noisy geometric attributes, (ii) attribute embeddings corresponding to each element, and (iii) timestep embeddings encoding the temporal context in the denoising schedule. These are concatenated along the element and feature dimensions, passed through multiple layers of multi-head self-attention, and ultimately mapped via a fully-connected output head to a noise residue prediction vector. The training objective minimizes a simplified mean-squared error between the model’s predicted noise and the true noise injected in the forward process, is defined recursively by the noise schedule and ϵ∼N(0,I). Substantial empirical evaluation demonstrates that Transformer-based diffusion layout models —henceforth referred to as LayoutDM —exhibit superior performance over GAN/VAE baselines on multiple public layout datasets. These include RICO (mobile UI screens), PubLayNet (scientific document layouts), Magazine (magazine spread layouts), COCO (natural scene composition layouts), and TextLogo3K (text logo bounding-box datasets). Experiments are evaluated using rigorous metrics : Fiducial Inception Distance (FID) : measuring the statistical distance between real and generated layouts via learned feature representations; Max Intersection Over Union (Max IoU) : quantifying geometric similarity; Overlap (total overlapping area of bounding boxes); and Alignment loss (average visual misalignment of elements). In implementation, LayoutDM typically employs eight Transformer layers (multi-head, usually eight heads per layer), learns all embedding and denoising parameters via Adam optimization at a small learning rate, and adopts batch sizes as large as hardware allows (1,024 or more). All training and inference are performed in PyTorch or PyTorch Lightning and utilize modern GPU accelerators (e.g., NVIDIA Quadro RTX 6000). Key ablation studies compare performance under different component configurations: e.g., the effect of omitting attribute embeddings, varying the number of Transformer layers/heads, ablation of timestep encoding, or substituting the backbone with a convolutional denoiser. Results confirm that all such features—especially self-attention across attribute-rich, unordered element sets, and conditioning on semantic attributes—are crucial for state-of-the-art layout generation.Moreover, the successes of diffusion-based layout models open broader research avenues: learning multi-modal mappings between textual descriptions and layout structures, joint co-design of content and visual style, and robust adaptation to new application domains such as game level design , responsive website layouts , or augmented reality scene annotation. This paper proposes a transformer-based diffusion model LayoutDM to address conditional layout generation. We introduce a purely transformer-based Layout Denoiser to model the diffusion reverse process. Benefitting from both DDPM and transformer, in comparison to existing methods, LayoutDM can generate high-quality generation with desired properties such as better diversity, faithful distribution coverage, and stationary training. Quantitative and qualitative results demonstrate that our model outperforms the state-of-the-art methods in terms of visual perceptual quality and diversity.Although our method shows impressive results in the conditional layout generation problem in comparison to existing methods, it still has limitations. For example, like other layout generation methods, our approach treats design elements as being on a single-layer canvas. This can not model a layout with multiple layers occluding each other. Our method also has no advantage over other generative models in generation speed because the generation of the diffusion model requires an iterative denoising process. We leave the solution to the above problems for future work.
In contemporary artificial intelligence research, the problem of layout (Layout) generation—the automated arrangement of visually distinct objects or functional elements within a designed frame—remains of paramount practical and theoretical significance. Layouts underpin a broad spectrum of applications ranging from magazine publishing and poster production to the design of complex user interfaces in mobile and web environments. A well-constructed layout not only directs user attention and guides reading order, but also encodes semantic relationships among visual or textual components, incorporates constraints dictated by content attributes, and adheres to general and domain-specific aesthetic principles. Despite rapid progress in machine learning, the manual construction of graphic layouts remains an arduous and resource-intensive process. Designers are required to orchestrate element positioning, sizing, and alignment in accordance with both objective communicative needs and subjective artistic judgments, a task that can become intractable as the complexity and variability of both content and user-specified constraints increase. This motivates the development of intelligent systems for automatic layout generation (Automatic Layout Generation), allowing efficient synthesis of designs that satisfy prescribed structural and semantic requirements. Early computational methods for layout synthesis often relied on the formalization of expert rules as handcrafted energy functions or procedural grammars, encoding a priori notions of spatial harmony and alignment into optimization objectives. However, such rule-based or heuristic-driven systems lack scalability and generality, particularly when confronted with data distributions that exhibit significant heterogeneity or intricate relationships between design eleents (Design Elements). They are also limited by their dependence on manual encoding of preferences—a process difficult to generalize across diverse domains such as editorial layout, digital ads, and mobile UI screens. Recent advances in deep generative models have revolutionized automatic layout generation by allowing systems to learn directly from large-scale, annotated datasets . Two especially influential approaches in this context are Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) . GANs are formulated as two-player zero-sum games between a generator that synthesizes candidate samples and a discriminator that assesses the realism of those samples relative to observed data. Such adversarial training fosters the creation of highly plausible, often photorealistic outputs. In contrast, VAEs employ probabilistic encodings and Bayesian inference to model latent variable distributions over input data, enabling diversity and smooth interpolation between generated layouts. Each method, however, exhibits distinct limitations in the context of layout synthesis . GANs are infamous for issues such as training instability , mode collapse , and incomplete distribution coverage ; these can impede their ability to accurately recover the rich diversity of real-world layouts, especially in visually cluttered or structurally complex domains. VAEs , while often easier to train and able to capture a broader range of plausible alternatives, can suffer from trade-offs between sharpness and diversity , with generated layouts sometimes lacking the visual fidelity required for professional design. More recently, research attention has shifted to a new class of generative models—the Diffusion Models (Diffusion Models)—notably represented by the Denoising Diffusion Probabilistic Model (DDPM) and related architectures. Diffusion models pose data generation as a progressive stochastic process, starting from pure Gaussian noise and iteratively denoising to synthesize high-quality structured outputs. The forward process gradually injects noise into training data according to a predefined noise schedule, while the reverse process (reverse process) employs a neural network (typically a U-Net ) to remove noise in discrete steps, reconstructing data samples that closely approximate the target distribution. Key advantages of diffusion models include the ability to achieve a stationary training objective (stationary training objective)—circumventing adversarial oscillation—as well as the capacity to faithfully cover the underlying data distribution . Moreover, these models afford strong sample diversity and permit straightforward conditioning on auxiliary information, making them theoretically attractive for the complex, multi-constraint setting of graphical layout generation . However, several key challenges must be addressed to adapt diffusion models to the specific structure of layout data . First, unlike grid-structured images or time-sequenced signals, a graphic layout typically comprises a variable-length, unordered set of elements, each described by both discrete (e.g., class label) and continuous (e.g., size, coordinates) attributes. Conventional convolutional neural networks (CNNs) , which excel at learning regular pixel grids but presume spatial locality, are generally ill-suited for processing such sets, especially when explicit modeling of relationships and attribute constraints is required. Second, the semantics of a layout are not solely a function of its individual components but are emergent from the interdependencies among elements, meaning that conditional generation demands sophisticated mechanisms for relational reasoning . Addressing these obstacles, contemporary research has advocated for the use of Transformer-based architectures (Transformer-based architectures) as the neural backbone in layout diffusion models. The self-attention mechanism intrinsic to Transformer layers is adept at modeling dependencies in unordered and variable-sized input sets, enabling efficient aggregation of global context and complex attribute interactions. Specifically, in the context of layout diffusion , a Transformer-based denoiser can predict noise vectors for each element token based on global information , facilitating accurate and flexible reverse diffusion even when layout cardinality varies significantly across samples. In the conditional generation setting, such architectures further allow for the inclusion of rich attribute embeddings —capturing category labels, style codes, and textual descriptors—so that the generated layouts respect user-controlled constraints and design intents. Crucially, because the arrangement of layout elements is unordered by nature, these models typically omit positional encoding (positional encoding), diverging from mainstream practices in natural language processing or image modeling where sequence order is semantically essential. To formally describe the layout modeling process, suppose that a layout instance L is a set of N elements, with each element ei represented as a tuple of continuous geometry gi (position, width, height) and discrete or textual attributes fi (e.g., “heading”, “button”, or “price”). For consistency and stability, geometric parameters are normalized onto a fixed unit hypercube. The generative apparatus—namely, a conditional denoising diffusion probabilistic model (conditional DDPM)—defines a forward process, which stepwise corrupts g by adding Gaussian noise, and a reverse process, which sequentially denoises g back to its original configuration, all conditioned on f. The conditional layout denoiser (cLayoutDenoiser) is realized as a sequence of stacked Transformer blocks . For each sample and each reverse diffusion time step t, the model receives as input (i) geometry embeddings , i.e., learned projections of noisy geometric attributes, (ii) attribute embeddings corresponding to each element, and (iii) timestep embeddings encoding the temporal context in the denoising schedule. These are concatenated along the element and feature dimensions, passed through multiple layers of multi-head self-attention, and ultimately mapped via a fully-connected output head to a noise residue prediction vector. The training objective minimizes a simplified mean-squared error between the model’s predicted noise and the true noise injected in the forward process, is defined recursively by the noise schedule and ϵ∼N(0,I). Substantial empirical evaluation demonstrates that Transformer-based diffusion layout models —henceforth referred to as LayoutDM —exhibit superior performance over GAN/VAE baselines on multiple public layout datasets. These include RICO (mobile UI screens), PubLayNet (scientific document layouts), Magazine (magazine spread layouts), COCO (natural scene composition layouts), and TextLogo3K (text logo bounding-box datasets). Experiments are evaluated using rigorous metrics : Fiducial Inception Distance (FID) : measuring the statistical distance between real and generated layouts via learned feature representations; Max Intersection Over Union (Max IoU) : quantifying geometric similarity; Overlap (total overlapping area of bounding boxes); and Alignment loss (average visual misalignment of elements). In implementation, LayoutDM typically employs eight Transformer layers (multi-head, usually eight heads per layer), learns all embedding and denoising parameters via Adam optimization at a small learning rate, and adopts batch sizes as large as hardware allows (1,024 or more). All training and inference are performed in PyTorch or PyTorch Lightning and utilize modern GPU accelerators (e.g., NVIDIA Quadro RTX 6000). Key ablation studies compare performance under different component configurations: e.g., the effect of omitting attribute embeddings, varying the number of Transformer layers/heads, ablation of timestep encoding, or substituting the backbone with a convolutional denoiser. Results confirm that all such features—especially self-attention across attribute-rich, unordered element sets, and conditioning on semantic attributes—are crucial for state-of-the-art layout generation.Moreover, the successes of diffusion-based layout models open broader research avenues: learning multi-modal mappings between textual descriptions and layout structures, joint co-design of content and visual style, and robust adaptation to new application domains such as game level design , responsive website layouts , or augmented reality scene annotation. This paper proposes a transformer-based diffusion model LayoutDM to address conditional layout generation. We introduce a purely transformer-based Layout Denoiser to model the diffusion reverse process. Benefitting from both DDPM and transformer, in comparison to existing methods, LayoutDM can generate high-quality generation with desired properties such as better diversity, faithful distribution coverage, and stationary training. Quantitative and qualitative results demonstrate that our model outperforms the state-of-the-art methods in terms of visual perceptual quality and diversity.Although our method shows impressive results in the conditional layout generation problem in comparison to existing methods, it still has limitations. For example, like other layout generation methods, our approach treats design elements as being on a single-layer canvas. This can not model a layout with multiple layers occluding each other. Our method also has no advantage over other generative models in generation speed because the generation of the diffusion model requires an iterative denoising process. We leave the solution to the above problems for future work.
考点1: 描述“归一化到固定单位超立方体”时,必须用“normalized onto a fixed unit hypercube”表达,不能生造新词。 考点2:“Fréchet Inception Distance”是一个专有名词缩写,学术论文中通常直接保留英文缩写,中文全称可以写作 Fréchet Inception 距离
2c2bb
学术论文
应用学科
179
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 中华人民共和国福建省厦门市中级人民法院民事判决书 (1997)厦民三初字第20号 原告:厦门赣东实业有限公司 住所:厦门市三湾路25号 法定代表人:陈志坚,总经理 原告:新加坡KALIO机械公司 住所:新加坡科技大道837号 法定代表人:金志河,该公司总裁 委托代理人:刘克添,福建省方正律师事务所律师 被告:三亚市安易实业有限公司 住所:海南省三亚市赤屿路23号 法定代表人:叶轩河,该公司总经理 委托代理人:许可东,福建远东律师事务所律师 原告厦门赣东实业有限公司(以下简称赣东公司)、新加坡KALIO机械公司(以下简称KALIO公司)与三亚市安易实业有限公司(以下简称安易公司)合同纠纷一案,本院受理后,依法组成合议庭,公开开庭进行了审理。原告赣东公司法定代表人陈志坚,KALIO公司委托代理人刘克添,被告安易公司委托代理人许可东到庭参加诉讼。本案现已审理终结。 原告厦门赣东公司、新加坡KALIO公司诉称: 1998年5月6日,赣东公司、KALIO公司共同作为乙方与作为甲方的安易公司签订了纺织机买卖合同。合同约定:甲方向乙方订购730部新加坡产开泰牌Z61型编织机;交货时间从1998年5月8日起至1999年1月8日止;价格按FOB(新加坡)每台20000元人民币;交货地点为厦门港;付款方式为机器运抵厦门港后付清全部货款;运输方法及费用负担:海运费用由甲方负担;违约责任,如单方违约,违约方必须向对方赔偿标准为未执行部分合同总额的10%的违约金。合同签订后,原告方按约定给被告发运了价值为900000元的全自动编织机及部分配件。安易公司陆续给付了原告机款500000元,现尚欠原告方机款400000元未付。另外,为履行合同,赣东公司为安易公司发运编织机已垫付运费5000元。 二原告提出如下诉讼请求: 1.请求法院判决被告向原告给付400000元及违约金90000元,运费5000元; 2.请求法院判决被告二向原告赔偿其经济损失共计100000元; 3.请求法院判决诉讼费用由被告承担。 二原告向法院提交了以下证据材料: 1.证据一为原被告签订的纺织机买卖合同。该合同由甲方安易公司加盖单位公章,法定代表人叶轩河签名,乙方赣东公司代理人陈志坚签名,KALIO公司加盖单位公章、代表人金志河签名。由此证明二原告与被告所签订的纺织机合同为有效合同,二原告与被告都应当按照合同约定履行各自义务。因此,被告方应给付拖欠的货款并承担相应的违约责任。 2.证据二为厦门海关进口关税专用缴款书及厦门边境贸易公司代理进口证明。证明二可证明:1998年5月10日,厦门边境贸易公司作为赣东公司的代理人,从新加坡进口了57台纺织机,赣东公司于1998年8月12日向厦门边境贸易公司交纳了7700元的纺织机的代理费、办证费、商检费、口岸费等。 3.证据三为在合同履行期间,安易公司的法定代表人叶轩河与赣东公司的委托代理人陈志坚的多次往来信件。由此证明:原告已按照上述购销合同实际履行,不存在任何违约的情形。 4.被告对于原告提交的三项证据都没有异议。 本法院认为:二原告与被告所签订的纺织机购销合同为有效合同,合同各方都应当按照合同约定履行各自义务。现被告方未履行付款义务,则被告方应给付拖欠的货款并承担违约责任。二原告要求被告给付400000元及违约金90000元,运费5000元的主张,本院予以支持;二原告要求被告赔偿100000元人民币损失的主张无事实依据,本院驳回请求。 依照《中华人民共和国经济合同法》第六条、第29条第1款、第31条、第32条、《中华人民共和国民法通则》第一百零六条、第61条第1款之规定,判决: 1.安易公司于本判决生效之日起十日内偿付赣东公司、KALIO公司纺织机及配件款400000元,运费5000元,并支付违约金90000元,合计495000元; 2.案件受理费20000元,由被告负担。 如不服本判决,可在判决书收到之日起十五日内向本院递交上诉状,并按对方当事人的人数提出副本,上诉于福建省高级人民法院。 审判长:刘国选 审判员:张俊中 审判员:马俊楠 日期:1999年3月3日 书记员:江南 (厦门市中级人民法院盖章)
中华人民共和国福建省厦门市中级人民法院民事判决书(1997)厦民三初字第20号原告:厦门赣东实业有限公司住所:厦门市三湾路25号法定代表人:陈志坚,总经理 原告:新加坡KALIO机械公司住所:新加坡科技大道837号法定代表人:金志河,该公司总裁委托代理人:刘克添,福建省方正律师事务所律师 被告:三亚市安易实业有限公司住所:海南省三亚市赤屿路23号法定代表人:叶轩河,该公司总经理委托代理人:许可东,福建远东律师事务所律师 原告厦门赣东实业有限公司(以下简称赣东公司)、新加坡KALIO机械公司(以下简称KALIO公司)与三亚市安易实业有限公司(以下简称安易公司)合同纠纷一案,本院受理后,依法组成合议庭,公开开庭进行了审理。原告赣东公司法定代表人陈志坚,KALIO公司委托代理人刘克添,被告安易公司委托代理人许可东到庭参加诉讼。本案现已审理终结。原告厦门赣东公司、新加坡KALIO公司诉称:1998年5月6日,赣东公司、KALIO公司共同作为乙方与作为甲方的安易公司签订了纺织机买卖合同。合同约定:甲方向乙方订购730部新加坡产开泰牌Z61型编织机;交货时间从1998年5月8日起至1999年1月8日止;价格按FOB(新加坡)每台20000元人民币;交货地点为厦门港;付款方式为机器运抵厦门港后付清全部货款;运输方法及费用负担:海运费用由甲方负担;违约责任,如单方违约,违约方必须向对方赔偿标准为未执行部分合同总额的10%的违约金。合同签订后,原告方按约定给被告发运了价值为900000元的全自动编织机及部分配件。安易公司陆续给付了原告机款500000元,现尚欠原告方机款400000元未付。另外,为履行合同,赣东公司为安易公司发运编织机已垫付运费5000元。 二原告提出如下诉讼请求:1.请求法院判决被告向原告给付400000元及违约金90000元,运费5000元;2.请求法院判决被告二向原告赔偿其经济损失共计100000元;3.请求法院判决诉讼费用由被告承担。 二原告向法院提交了以下证据材料:1.证据一为原被告签订的纺织机买卖合同。该合同由甲方安易公司加盖单位公章,法定代表人叶轩河签名,乙方赣东公司代理人陈志坚签名,KALIO公司加盖单位公章、代表人金志河签名。由此证明二原告与被告所签订的纺织机合同为有效合同,二原告与被告都应当按照合同约定履行各自义务。因此,被告方应给付拖欠的货款并承担相应的违约责任。2.证据二为厦门海关进口关税专用缴款书及厦门边境贸易公司代理进口证明。证明二可证明:1998年5月10日,厦门边境贸易公司作为赣东公司的代理人,从新加坡进口了57台纺织机,赣东公司于1998年8月12日向厦门边境贸易公司交纳了7700元的纺织机的代理费、办证费、商检费、口岸费等。3.证据三为在合同履行期间,安易公司的法定代表人叶轩河与赣东公司的委托代理人陈志坚的多次往来信件。由此证明:原告已按照上述购销合同实际履行,不存在任何违约的情形。4.被告对于原告提交的三项证据都没有异议。 本法院认为:二原告与被告所签订的纺织机购销合同为有效合同,合同各方都应当按照合同约定履行各自义务。现被告方未履行付款义务,则被告方应给付拖欠的货款并承担违约责任。二原告要求被告给付400000元及违约金90000元,运费5000元的主张,本院予以支持;二原告要求被告赔偿100000元人民币损失的主张无事实依据,本院驳回请求。依照《中华人民共和国经济合同法》第六条、第29条第1款、第31条、第32条、《中华人民共和国民法通则》第一百零六条、第61条第1款之规定,判决:1.安易公司于本判决生效之日起十日内偿付赣东公司、KALIO公司纺织机及配件款400000元,运费5000元,并支付违约金90000元,合计495000元;2.案件受理费20000元,由被告负担。如不服本判决,可在判决书收到之日起十五日内向本院递交上诉状,并按对方当事人的人数提出副本,上诉于福建省高级人民法院。审判长:刘国选审判员:张俊中审判员:马俊楠日期:1999年3月3日书记员:江南(厦门市中级人民法院盖章)
考点1:“民事判决书”推荐译为“Civil Judgment” 考点2:“住所”推荐译为“domicile” 考点3:“委托代理人”推荐译为“authorized counsel” 考点4:“合议庭”推荐译为“collegial panel” 考点5:“公开开庭”推荐译为“opened a court session publicly” 考点6:“本案现已审理终结”推荐译为“the trial has now concluded” 考点7:“海运费用”推荐译为“‌ocean freight fee” 考点8:“违约金”推荐译为“liquidated damages” 考点9:“提出如下诉讼请求”推荐译为“put forward the following claims” 考点10:“诉讼费用”推荐译为“the court cost” 考点11:“进口关税缴款书”推荐译为“Payment Certificate for Import Duties” 考点12:“交纳口岸费”推荐译为“pay for port management” 考点13:“依照……的规定,判决:”推荐译为“in accordance with the stipulation of……,it orders as follows” 考点14:“第61条第1款”推荐译为“Article 61 Paragraph 1” 考点15:“驳回请求”推荐译为“dismiss the claims” 考点16:“向本院递交上诉状,上诉于福建省高级人民法院”推荐译为“bring the appeal to Fujian High People's Court via this court”
2cb64
垂类场景
法律
99
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 三、基于“流量池”的叠加推荐    与算法的研究日益成为热点类似,流量池也是近段时间以来频繁出现的热门词汇。此前,在新媒体研究领域里,流量思维是较为集中的研究热点,甚至有学者指出,“得流量者得天下”。流量池思维则是要获取流量并通过存储、运营和发掘等手段,进行信息的再传播,以期获得更多的流量。流量思维和流量池思维最大的区别就是流量获取之后的下一步社会行为,后者更强调如何用一批老用户找到更多新的用户,而流量思维更多的是首轮传播的效果评价。通常来讲,在内容流量池表现较好的视频内容往往会进入叠加推荐的行列,从而获得更多的阅读量和点赞。叠加推荐是以内容的综合权重做评估标准,综合权重的关键指标有完播率、点赞量、评论量、转发量,且每个元素所具有的影响权重又互不相同,当达到一定量级,平台就会以大数据算法和人工运营相结合的机制进行不断的推荐。比如,当用户发布一条视频时,平台会自动将其分配到一个流量池当中,分发到一定数量级的用户的推荐界面,然后通过统计该视频的播放效果,形成一个加权分数,转发量、评论量、点赞量的权重依次递减,分数越高则获得叠加推荐的机会越大,播放效果好的视频会再次加入流量池进行更大范围的分发,而表现较差的视频则失去了被推荐的机会,沉淀至流量池底部。第二次推荐又获得比较好的反馈则进入下一轮的推荐,从而获得更大规模的推荐。流量池推荐带来了更加明显的“马太效应”,优质的视频被反复推荐,获得更大的积累优势,而在第一波流量池推送中“逊色”的视频则失去了更大规模被推荐的机会。同时,由于这类算法更加基于多重用户受众的实际行为分析,所以经常会有大量级播放次数的视频出现。与前两个算法推荐相比,流量池推荐的视频随机性更强,其推荐的法则并非主要着眼于视频内容,而更多的是通过用户的反馈进行推荐。换言之,流量池推荐更注重视频传播效果的评价而非内容生产的优劣。这也使得一些内容并不那么优质的视频利用算法推荐的漏洞或不足,能够取得很高的评价分数,从而被大量级推荐,迅速“蹿红”。 四、建议与对策    依靠强大的算法推荐,抖音已经在目前短视频白热化的角逐中逐渐占据上风。但是,完全依靠协同过滤和精准分发的单纯算法推荐不能够充分适应目前快速发展的受众需求,平台需要进一步完善更多维度的算法推荐系统。采用更加多元和开放的算法,将会更加合理地促进优质视频内容的传播。因此,可以进一步加强对抖音用户使用行为的数据挖掘工作,引入满意度、有用性等其他考量维度,优化其产品内容评价体系分层分类,避免加剧“茧房效应”,适当调整兴趣内容与其他内容的分发比例;进一步提高平台的识别能力,对所谓的“网红”短视频内容加强甄别,提高原创视频的推广力度,加强对相似或较为雷同的视频进行过滤审核,避免“同质化”内容高频出现;积极引入或增加人工审核的机制,在审核过程中提高人工排查的参与程度,逐渐树立传播审核过程中“人”的参与意识;积极鼓励“PGC”的产品生产,对于具有优质视频生产能力的用户给予鼓励,通过身份认证、延长视频时限等方式给予支持,提升全平台的视频制作水平;始终坚持“内容为王”的运营和管理理念,依靠优质的平台内容增加用户黏度和吸引新用户,就一些具有明显“哗众取宠”和恶搞的内容,引入投诉和其他负面评价机制,进一步净化网络空间。
三、基于“流量池”的叠加推荐    与算法的研究日益成为热点类似,流量池也是近段时间以来频繁出现的热门词汇。此前,在新媒体研究领域里,流量思维是较为集中的研究热点,甚至有学者指出,“得流量者得天下”。流量池思维则是要获取流量并通过存储、运营和发掘等手段,进行信息的再传播,以期获得更多的流量。流量思维和流量池思维最大的区别就是流量获取之后的下一步社会行为,后者更强调如何用一批老用户找到更多新的用户,而流量思维更多的是首轮传播的效果评价。通常来讲,在内容流量池表现较好的视频内容往往会进入叠加推荐的行列,从而获得更多的阅读量和点赞。叠加推荐是以内容的综合权重做评估标准,综合权重的关键指标有完播率、点赞量、评论量、转发量,且每个元素所具有的影响权重又互不相同,当达到一定量级,平台就会以大数据算法和人工运营相结合的机制进行不断的推荐。比如,当用户发布一条视频时,平台会自动将其分配到一个流量池当中,分发到一定数量级的用户的推荐界面,然后通过统计该视频的播放效果,形成一个加权分数,转发量、评论量、点赞量的权重依次递减,分数越高则获得叠加推荐的机会越大,播放效果好的视频会再次加入流量池进行更大范围的分发,而表现较差的视频则失去了被推荐的机会,沉淀至流量池底部。第二次推荐又获得比较好的反馈则进入下一轮的推荐,从而获得更大规模的推荐。流量池推荐带来了更加明显的“马太效应”,优质的视频被反复推荐,获得更大的积累优势,而在第一波流量池推送中“逊色”的视频则失去了更大规模被推荐的机会。同时,由于这类算法更加基于多重用户受众的实际行为分析,所以经常会有大量级播放次数的视频出现。与前两个算法推荐相比,流量池推荐的视频随机性更强,其推荐的法则并非主要着眼于视频内容,而更多的是通过用户的反馈进行推荐。换言之,流量池推荐更注重视频传播效果的评价而非内容生产的优劣。这也使得一些内容并不那么优质的视频利用算法推荐的漏洞或不足,能够取得很高的评价分数,从而被大量级推荐,迅速“蹿红”。 四、建议与对策    依靠强大的算法推荐,抖音已经在目前短视频白热化的角逐中逐渐占据上风。但是,完全依靠协同过滤和精准分发的单纯算法推荐不能够充分适应目前快速发展的受众需求,平台需要进一步完善更多维度的算法推荐系统。采用更加多元和开放的算法,将会更加合理地促进优质视频内容的传播。因此,可以进一步加强对抖音用户使用行为的数据挖掘工作,引入满意度、有用性等其他考量维度,优化其产品内容评价体系分层分类,避免加剧“茧房效应”,适当调整兴趣内容与其他内容的分发比例;进一步提高平台的识别能力,对所谓的“网红”短视频内容加强甄别,提高原创视频的推广力度,加强对相似或较为雷同的视频进行过滤审核,避免“同质化”内容高频出现;积极引入或增加人工审核的机制,在审核过程中提高人工排查的参与程度,逐渐树立传播审核过程中“人”的参与意识;积极鼓励“PGC”的产品生产,对于具有优质视频生产能力的用户给予鼓励,通过身份认证、延长视频时限等方式给予支持,提升全平台的视频制作水平;始终坚持“内容为王”的运营和管理理念,依靠优质的平台内容增加用户黏度和吸引新用户,就一些具有明显“哗众取宠”和恶搞的内容,引入投诉和其他负面评价机制,进一步净化网络空间。
考点1:“叠加推荐”推荐译为“Tiered Recommendations” 考点2:“得流量者得天下”推荐译为“Win the Eyeballs, Win it All.” 考点3:“热门词汇”推荐译为“buzzword” 考点4:“综合权重的关键指标”推荐译为“key performance metrics” 考点5:“播放效果”推荐译为“user engagement” 考点6:“蹿红”推荐译为“go viral” 考点7:“白热化的角逐”推荐译为“the fiercely competitive landscape” 考点8:“茧房效应”推荐译为““filter bubble” effect” 考点9:““同质化”内容”推荐译为“content homogenization” 考点10:“恶搞”推荐译为“maliciously sensational” 考点11:“内容为王”推荐译为“Content is king.” 考点12:“抖音”应译为“Douyin”,译为“Tiktok”算错误
2d22f
学术论文
社会科学
14
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 27日,高德地图宣布在韩国正式上线打车功能,中国游客无需更换手机卡或下载新应用,就能直接呼叫韩国本地出租车,支持中文界面叫车、跨语言沟通、实时导航和主流电子支付,加之此前就已在当地上线的地图导航服务,实现“一个APP便捷出行”。据悉,高德地图打车功能目前已覆盖首尔、釜山等韩国主要旅游城市。 韩国是中国游客的重要旅游目的地之一,公开数据显示,2024年赴韩中国游客达460万人次,较2023年增长翻番。针对出境游中的语言障碍、支付门槛等痛点,高德地图基于全球领先的基础服务能力与韩国本土出行服务商合作,将国内成熟的出行服务模式延展至海外,让游客在韩国也能享受到便捷、高效的线上叫车服务。 据了解,中国游客在韩旅游期间,不仅可以在高德地图上直接使用中文地址一键呼叫韩国出租车,通过APP内置通信联络司机时,AI翻译能够实时转换司乘间的中韩对话,消除语言障碍。支付环节则直连支付宝、微信等跨境电子钱包,汇率自动结算,有效解决跨境支付难题。 6月12日,阿里巴巴集团旗下高德地图在日本正式上线打车功能,覆盖东京、大阪、名古屋、横滨、京都等日本热门旅游城市。这是继韩国市场之后,高德再次通过“技术创新+本地合作”的模式,持续完善海外出行服务网络,为用户提供“一个APP便捷出行”的旅行体验。据悉,高德地图用户可通过中文界面直接呼叫日本当地出租车,可以使用支付宝、微信等跨境支付。 高德介绍,通过与日本出租车头部平台“GO”合作,将国内成熟的出行服务模式延伸至日本市场,成为高德国际化战略中的又一重要动作。即日起,中国内地及港澳地区用户无需更换手机卡或下载新应用,即可通过高德地图中文界面直接呼叫日本当地出租车,实现中文地址一键叫车、中日对话实时翻译,以及支付宝、微信跨境支付,有效解决了语言障碍和支付难题。结合此前就已上线的实时导航、景点查询等功能,高德进一步完善了一体化出行服务能力,以更好满足游客在异国他乡的出行需求。 高德打车总裁王桂馨表示,此次在日本上线打车服务,是公司产品走向全球的重要一步。“我们始终坚持以用户需求为核心,致力于将便捷可靠的出行服务带到世界各地,通过本地化的产品创新和生态协同,为中国以及海外用户的全球出行,创造更优质的体验。” 目前,高德地图在日本的打车服务已基于“GO”的覆盖网络,延伸至日本各大热门旅游城市。暑期出境游高峰将至,该功能上线后能为游客提供更便捷、丰富的数字化出行选择,让用户在日本旅游时也能享受与国内同等效率的叫车服务。
27日,高德地图宣布在韩国正式上线打车功能,中国游客无需更换手机卡或下载新应用,就能直接呼叫韩国本地出租车,支持中文界面叫车、跨语言沟通、实时导航和主流电子支付,加之此前就已在当地上线的地图导航服务,实现“一个APP便捷出行”。据悉,高德地图打车功能目前已覆盖首尔、釜山等韩国主要旅游城市。 韩国是中国游客的重要旅游目的地之一,公开数据显示,2024年赴韩中国游客达460万人次,较2023年增长翻番。针对出境游中的语言障碍、支付门槛等痛点,高德地图基于全球领先的基础服务能力与韩国本土出行服务商合作,将国内成熟的出行服务模式延展至海外,让游客在韩国也能享受到便捷、高效的线上叫车服务。 据了解,中国游客在韩旅游期间,不仅可以在高德地图上直接使用中文地址一键呼叫韩国出租车,通过APP内置通信联络司机时,AI翻译能够实时转换司乘间的中韩对话,消除语言障碍。支付环节则直连支付宝、微信等跨境电子钱包,汇率自动结算,有效解决跨境支付难题。 6月12日,阿里巴巴集团旗下高德地图在日本正式上线打车功能,覆盖东京、大阪、名古屋、横滨、京都等日本热门旅游城市。这是继韩国市场之后,高德再次通过“技术创新+本地合作”的模式,持续完善海外出行服务网络,为用户提供“一个APP便捷出行”的旅行体验。据悉,高德地图用户可通过中文界面直接呼叫日本当地出租车,可以使用支付宝、微信等跨境支付。 高德介绍,通过与日本出租车头部平台“GO”合作,将国内成熟的出行服务模式延伸至日本市场,成为高德国际化战略中的又一重要动作。即日起,中国内地及港澳地区用户无需更换手机卡或下载新应用,即可通过高德地图中文界面直接呼叫日本当地出租车,实现中文地址一键叫车、中日对话实时翻译,以及支付宝、微信跨境支付,有效解决了语言障碍和支付难题。结合此前就已上线的实时导航、景点查询等功能,高德进一步完善了一体化出行服务能力,以更好满足游客在异国他乡的出行需求。 高德打车总裁王桂馨表示,此次在日本上线打车服务,是公司产品走向全球的重要一步。“我们始终坚持以用户需求为核心,致力于将便捷可靠的出行服务带到世界各地,通过本地化的产品创新和生态协同,为中国以及海外用户的全球出行,创造更优质的体验。” 目前,高德地图在日本的打车服务已基于“GO”的覆盖网络,延伸至日本各大热门旅游城市。暑期出境游高峰将至,该功能上线后能为游客提供更便捷、丰富的数字化出行选择,让用户在日本旅游时也能享受与国内同等效率的叫车服务。
考点1:“打车功能”推荐译为“ride-hailing feature/service; ride-hailing function” 考点2:“基础服务能力”推荐译为“fundamental service capabilities; core service infrastructure” 考点3:“ 一键呼叫”推荐译为“one-tap hailing; one-click hailing;” 考点4:“头部平台”推荐译为“leading platform; top platform” 考点5:“一体化出行服务”推荐译为“ integrated mobility services” 考点6:“暑期出境游高峰”推荐译为“the summer peak season for outbound travel; the peak of summer outbound tourism”
2ec7e
新闻资讯
新闻报道
6
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: **考查重点:语文素养、时代意识、家国情怀**   2025年高考语文共有5套试卷。其中,教育部教育考试院命制全国一卷和全国二卷的2道作文题,上海、天津各命制1道,北京命制2道和1篇微写作。   教育部教育考试院有关负责人表示,今年的高考作文题坚持价值引领,紧扣时代脉搏,彰显学科特色,重点考查语文素养、时代意识与家国情怀。   今年是中国人民抗日战争暨世界反法西斯战争胜利80周年。全国一卷作文题选取了三段和抗战有关的材料,将考生的视线拉回到那段浴血奋战的峥嵘岁月,引发对民族精神传承的思考。   此外,全国二卷作文题鼓励考生敢于有梦、勇于追梦;天津卷以大家熟悉的车轮、辐条等为意象,推动考生思考向心力、凝聚力对个体成长、事业发展的重要意义。   “坚持立德树人,这些试题注重引导考生树立坚定理想信念,厚植家国情怀,可以激发新时代青年勇于担当的力量。”中国教育在线总编辑陈志文说。 贴近实际、贴近生活,是今年高考作文题的另一亮点。   例如,上海卷作文题围绕专业文章、通俗文章、传世文章展开,与当下的文化生活紧密相连;北京卷则从运动员中长跑时会出现的“极点”反应,以及日常生活中随处可见的比赛记分牌、车站电子时刻表等角度出发命题。   “今年的上海高考作文题与考生的日常经验比较贴合。”华东师范大学教授胡晓明分析表示,“当前,传播手段非常丰富,许多网络文章被广泛转发,不少经典作品也得到更广泛的传播。从这个角度来看,学生若对当前文化传播有一定程度了解,便可以打开思辨空间,写出言之有物的文章。”   在专家看来,这些题目的材料指向既可以是现实的,又可以是想象的,由现实展开想象,由想象观照现实,考生可以从自身生活经验谈起,切入的角度很多。 **折射导向:减少机械刷题、鼓励创新表达、促进全面发展**   每年的高考作文题,都可以折射一段时间以来语文教育的发展和趋势。多位受访专家和一线教师认为,今年的作文题同样为未来一段时间的语文教育提供了新参考和新导向。   衔接教学,促进课堂提质——   在有关专家看来,以作文题为代表的2025年高考语文试题注重与教材的关联,致力于引导一线教学用好统编教材的同时学以致用。   例如,全国一卷作文题的第二则材料选自艾青的名篇《我爱这土地》,这首诗正是统编教材九年级上册的篇目之一。专家认为,这类试题要求学生在重视课堂学习的同时,做到对课内教学内容融会贯通,扭转机械刷题、以练代学的不良习惯;也要求教师在课堂上开展深度教学,提升课堂教学质量。   以人为本,鼓励创新表达——   纵观近年来的高考作文题,不少专家发现,越来越多题目不在审题立意上设置门槛,而是更加注重激发考生活跃的想象力和创造力,打开审题立意的多维空间。  “从中也反映出语文教育的趋势是引导学生创新思维,大胆联想和想象。”李奇说,“这启示一线教师在教育教学过程中引导学生从苦练苦背型学习,转向探究运用型学习,从而提升学生的思维整合力、思考纵深度。”   提升素养,促进全面发展——   从古代诗歌到现代诗歌,从观察生活到深入论证,专家表示,今年的作文题对考生的综合素养提出了新要求,引导学生不断提升阅读能力、思考能力和表达能力,实现全面发展。   在柳州高级中学语文教研组长朱秋清看来,今年的高考作文题对学生综合能力要求较高。“未来的教学实践中,我们既要帮助学生打牢基础,也要引导他们更加关注现实、思考人生,充分调动他们在学习和生活中积累的经验,让学生可以在课堂教学、日常阅读、历史感悟和生活体验中实现更加全面的发展。”
**考查重点:语文素养、时代意识、家国情怀**   2025年高考语文共有5套试卷。其中,教育部教育考试院命制全国一卷和全国二卷的2道作文题,上海、天津各命制1道,北京命制2道和1篇微写作。   教育部教育考试院有关负责人表示,今年的高考作文题坚持价值引领,紧扣时代脉搏,彰显学科特色,重点考查语文素养、时代意识与家国情怀。   今年是中国人民抗日战争暨世界反法西斯战争胜利80周年。全国一卷作文题选取了三段和抗战有关的材料,将考生的视线拉回到那段浴血奋战的峥嵘岁月,引发对民族精神传承的思考。   此外,全国二卷作文题鼓励考生敢于有梦、勇于追梦;天津卷以大家熟悉的车轮、辐条等为意象,推动考生思考向心力、凝聚力对个体成长、事业发展的重要意义。   “坚持立德树人,这些试题注重引导考生树立坚定理想信念,厚植家国情怀,可以激发新时代青年勇于担当的力量。”中国教育在线总编辑陈志文说。 贴近实际、贴近生活,是今年高考作文题的另一亮点。   例如,上海卷作文题围绕专业文章、通俗文章、传世文章展开,与当下的文化生活紧密相连;北京卷则从运动员中长跑时会出现的“极点”反应,以及日常生活中随处可见的比赛记分牌、车站电子时刻表等角度出发命题。   “今年的上海高考作文题与考生的日常经验比较贴合。”华东师范大学教授胡晓明分析表示,“当前,传播手段非常丰富,许多网络文章被广泛转发,不少经典作品也得到更广泛的传播。从这个角度来看,学生若对当前文化传播有一定程度了解,便可以打开思辨空间,写出言之有物的文章。”   在专家看来,这些题目的材料指向既可以是现实的,又可以是想象的,由现实展开想象,由想象观照现实,考生可以从自身生活经验谈起,切入的角度很多。 **折射导向:减少机械刷题、鼓励创新表达、促进全面发展**   每年的高考作文题,都可以折射一段时间以来语文教育的发展和趋势。多位受访专家和一线教师认为,今年的作文题同样为未来一段时间的语文教育提供了新参考和新导向。   衔接教学,促进课堂提质——   在有关专家看来,以作文题为代表的2025年高考语文试题注重与教材的关联,致力于引导一线教学用好统编教材的同时学以致用。   例如,全国一卷作文题的第二则材料选自艾青的名篇《我爱这土地》,这首诗正是统编教材九年级上册的篇目之一。专家认为,这类试题要求学生在重视课堂学习的同时,做到对课内教学内容融会贯通,扭转机械刷题、以练代学的不良习惯;也要求教师在课堂上开展深度教学,提升课堂教学质量。   以人为本,鼓励创新表达——   纵观近年来的高考作文题,不少专家发现,越来越多题目不在审题立意上设置门槛,而是更加注重激发考生活跃的想象力和创造力,打开审题立意的多维空间。  “从中也反映出语文教育的趋势是引导学生创新思维,大胆联想和想象。”李奇说,“这启示一线教师在教育教学过程中引导学生从苦练苦背型学习,转向探究运用型学习,从而提升学生的思维整合力、思考纵深度。”   提升素养,促进全面发展——   从古代诗歌到现代诗歌,从观察生活到深入论证,专家表示,今年的作文题对考生的综合素养提出了新要求,引导学生不断提升阅读能力、思考能力和表达能力,实现全面发展。   在柳州高级中学语文教研组长朱秋清看来,今年的高考作文题对学生综合能力要求较高。“未来的教学实践中,我们既要帮助学生打牢基础,也要引导他们更加关注现实、思考人生,充分调动他们在学习和生活中积累的经验,让学生可以在课堂教学、日常阅读、历史感悟和生活体验中实现更加全面的发展。”
考点1:“高考”推荐译为 National College Entrance Examination 考点2:“时代意识”推荐译为 awareness of the times / sense of the times,不可译为timing awareness。 考点3:“家国情怀”推荐译为 patriotic sentiment / devotion to the nation and family,不可误译为nationalism或emotional attachment to country。 考点4:“命制”推荐译为 developed/set/designed (by the Examination Authority) 考点5:“价值引领”推荐译为 value guidance / value-oriented approach,注意体现政策导向性质 考点6:“紧扣时代脉搏”推荐译为 keep in step with the pulse of the times,为文化性表达,避免直译如close to the beat of the times。 考点7:“彰显学科特色”推荐译为 highlight the discipline-specific features 考点8:“峥嵘岁月”推荐译为 tumultuous years / the trying days of war 考点9:“民族精神传承”推荐译为 inheritance of national spirit / passing down the national ethos 考点10:“立德树人”推荐译为 fostering virtue through education,为教育核心术语 考点11:“负责人” 不可译为 “responsible person/officer”
2ff04
新闻资讯
新闻报道
28
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Such a requalification of the anthropological agenda was what Tania Stolze Lima and I wanted to contribute to when we proposed the concept of Amerindian perspectivism as the reconfiguration of a complex of ideas and practices whose power of intellectual disturbance has never been sufficiently appreciated (even if they found the word relevant) by Americanists, despite its vast diffusion in the New World To this we added the synoptic concept of multinaturalism, which presented Amerindian thought as an unsuspected partner, a dark precursor if you will, of certain contemporary philosophical programs, like those developing around theories of possible worlds, others that refuse to operate within the vicious dichotomies of modernity, or still others that, having registered the end of the hegemony of the kind of critique that demands an epistemological response to every ontological question, are slowly defining new lines of flight for thought under the rallying cries of transcendental empiricism and speculative realism. The two concepts emerged following an analysis of the cosmological presuppositions of "the metaphysics of predation" evoked in the last chapter. We found that this metaphysics, as can be deduced from Lévi-Strauss' summary of it, reaches its highest expression in the strong speculative yield of those indigenous categories denoting matrimonial alliance, phenomena that I translated with yet another concept: virtual affinity.!" Virtual affinity is the schematism characteristic of what Deleuze would have called the "Other-structure"' of Amerindian worlds and is indelibly marked by cannibalism, which is an omnipresent motif in their inhabitants' relational imagination. Interspecific perspectivism, ontological multinaturalism and cannibal alterity thus form the three aspects of an indigenous alter-anthropology that is the symmetrical and reverse transformation of Occidental anthropolo-gy—as symmetrical in Latour's sense as it is reverse in the sense of Wagner' "reverse anthropology." By drawing this triangle, we can enter into the orbit of one of the philosophies of "the exotic peoples" that Lévi-Strauss opposed to ours and attempt, in other words, to realize something of the imposing program outlined in the fourth chapter, "Geophilosophy," of What Is Philosophy?.. even if it will be at the price—but one we should always be ready to payof a certain methodological imprecision and intentional ambiguity. Our work's perfectly contingent point of departure was the sudden perception of a resonance between the results of our research on Amazonian cosmopolitics-on its notion of a perspectivist multiplicity intrinsic to the real-and a well-known parable on the subject of the conquest of the Americans recounted by Lévi-Strauss in Race and History: In the Greater Antilles, some years after the discovery of America, while the Spaniards sent out investigating commissions to ascertain whether or not the natives had a soul, the latter were engaged in the drowning of white prisoners in order to verify, through prolonged watching, whether or not their corpses were subject to putrification. (L.-S. 1978[1952): 329) In this conflict between the two anthropologies, the author perceived a baroque allegory of the fact that one of the typical manifestations of human nature is the negation of its own generality. A kind of congenital avarice preventing the extension of the predicates of humanity to the species as a whole appears to be one of its predicates. In sum, ethnocentrism could be said to be like good sense, of which perhaps it is just the apperceptive moment: the best distributed thing in the world. The format of the lesson is familiar, but that does not lessen its sting. Overestimating one's own humanity to the detriment of the contemptible other's reveals one's deep resemblance with it. Since the other of the Same (of the European) shows itself to be the same as the Other's other (of the indigenous), the Same ends up unwittingly showing itself to be the same as the Other. The anecdote fascinated Lévi-Strauss enough for him to repeat it in Iristes Tropiques. But there he added a supplementary, ironic twist, this time noting a difference (rather than this re-semblance) between the parties. While the Europeans relied on the social sciences in their investigations of the humanity of the other, the Indians placed their faith in the natural sciences; and where the former proclaimed the Indians to be animals, the latter were content to suspect the others might be gods. "Both attitudes show equal ignorance," Lévi-Strauss concluded, "but the Indian's behavior certainly had greater dignity" (1992: 76). If this is really how things transpired,' it forces us to conclude that, despite being just as ignorant on the subject of the other, the other of the Other was not exactly the same as the other of the Same. We could even say that it was its exact opposite, if not for the fact that the relation between these two others of humanity-animality and divinity—is conceived in indigenous worlds in completely different terms than those we have inherited from Christianity. The rhetorical contrast Lévi-Strauss draws succeeds because it appeals to our cosmological hierarchies rather than those of the Taino. In any case, consideration of this disequilibrium was what led us to the hypothesis that Amerindian ontological regimes diverge from those widespread in the West precisely with regard to the inverse semiotic functions they respectively attribute to soul and body. The marked dimension for the Spanish was the soul, whereas the Indian emphasized the body. The Europeans never doubted that the Indians had bodies--animals have them too-and the Indians in turn never doubted that the Europeans had souls, since animals and the ghosts of the dead do as well. Thus the Europeans' ethnocentrism consisted in doubting that the body of the other contained a soul formally similar to the one inhabiting their own bodies, while the ethnocentrism of the Indians, on the contrary, entailed doubting that the others' souls or spirits could possess a body materially similar to theirs. • In the semiotic terms of Roy Wagner, a Melanesianist who will quickly reveal himself to be a crucial intercessor in the theory of Amerindian perspectivism, the body belongs to the innate or spontaneous dimension of European ontology ("nature"), which is the counter-invented result of an operation of conventionalist symbolization, while the soul would be the constructed dimension, the fruit of a "differentiating" symbol-ization that "specifies and renders concrete the conventional world by tracing radical distinctions and concretizing the singular individuals of this world" (Wagner 1981: 42). In indigenous worlds, on the contrary, the soul "is experienced as ... a manifestation of the conventional order implicit in everything" and "sums up the ways in which its possessor is similar to others, over and above the ways in which he differs from them" (Wagner 1981: 94); the body, on the contrary, belongs to the sphere of what comes from the responsibility of agents and is one of the fundamental figures of something that has to be constructed against a universal and innate ground of an "immanent humanity" (Wagner 1981: 86-9), In short, European praxis consists in "making souls" (and differentiating cultures) on the basis of a given corporeal-material ground nature while indigenous praxis consists in "making bodies" (and differentiating species) on the basis of a socio-spiritual continuum, itself also given ... but in myth, as we will see. Wagner's conceptually dense and quite original theoretical system resists didactic summary; thus we request that the reader directly engage its most elegant and realized presentation in The Invention of Culture. Grosso modo, the Wagnerian semiotic can be said to be a theory of human and nonhuman practice conceived as exhaustively consisting in the recipro-cal, recursive operation of two modes of symbolization: (1) a collectiv-izing, conventional (or literal) symbolism where signs are organized in standardized contexts (semantic domains, formal languages, etc.) to the extent that they are opposed to a heterogeneous plane of "referents". that is, they are seen as symbolizing something other than themselves; and (2) a differentiating, inventive (or figurative) mode in which the world of phenomena represented by conventional symbolization is understood to be constituted by "symbols representing themselves," that is, events that simultaneously manifest as symbols and referents, thereby dissolving the conventional contrast. It should be observed, first of all, that the world of referents or the "real" is defined here as a semiotic effect: what is other to a sign is another sign having the singular capacity of "representing itself." The mode of existence of actual entities qua events or occasions is a tautegory. It should be stressed that the contrast between the two modes is itself the result of a conventionalist operation (and perception): the distinction between invention and convention is itself conventional, but at the same time every convention is produced through a counter-invention. The contrast is thus intrinsically recursive, especially if we understand that human cultures are fundamentally in conflict over the mode of symbolization they (conventionally) privilege as an element appropriated for action or invention, in reserving to the other the function of the "given." Cultures, human macrosystems of conventions, are distinguished by what they define as belonging to the sphere of the responsibilities of agents-the mode of the constructed-and by what belongs (because it is counter-constructed as belonging) to the world of the given or non-constructed. The core of any and every set of cultural conventions is a simple distinction as to what kind of contexts- the nonconventionalized ones or those of convention itself- are to be deliberately articulated in the course of human action, and what kind of contexts are to be counter-invented as "motivation" under the conventional mask of "the given" or "the innate." Of course I...] there are only two possibilities: a people who deliberately differentiate as the form of their action will invariably counter-invent a motivating collectivity as "innate," and a people who deliberately collectivize will counter-invent a motivating differentiation in this way. (Wagner 1981: 51)
Such a requalification of the anthropological agenda was what Tania Stolze Lima and I wanted to contribute to when we proposed the concept of Amerindian perspectivism as the reconfiguration of a complex of ideas and practices whose power of intellectual disturbance has never been sufficiently appreciated (even if they found the word relevant) by Americanists, despite its vast diffusion in the New World To this we added the synoptic concept of multinaturalism, which presented Amerindian thought as an unsuspected partner, a dark precursor if you will, of certain contemporary philosophical programs, like those developing around theories of possible worlds, others that refuse to operate within the vicious dichotomies of modernity, or still others that, having registered the end of the hegemony of the kind of critique that demands an epistemological response to every ontological question, are slowly defining new lines of flight for thought under the rallying cries of transcendental empiricism and speculative realism. The two concepts emerged following an analysis of the cosmological presuppositions of "the metaphysics of predation" evoked in the last chapter. We found that this metaphysics, as can be deduced from Lévi-Strauss' summary of it, reaches its highest expression in the strong speculative yield of those indigenous categories denoting matrimonial alliance, phenomena that I translated with yet another concept: virtual affinity.!" Virtual affinity is the schematism characteristic of what Deleuze would have called the "Other-structure"' of Amerindian worlds and is indelibly marked by cannibalism, which is an omnipresent motif in their inhabitants' relational imagination. Interspecific perspectivism, ontological multinaturalism and cannibal alterity thus form the three aspects of an indigenous alter-anthropology that is the symmetrical and reverse transformation of Occidental anthropolo-gy—as symmetrical in Latour's sense as it is reverse in the sense of Wagner' "reverse anthropology." By drawing this triangle, we can enter into the orbit of one of the philosophies of "the exotic peoples" that Lévi-Strauss opposed to ours and attempt, in other words, to realize something of the imposing program outlined in the fourth chapter, "Geophilosophy," of What Is Philosophy?.. even if it will be at the price—but one we should always be ready to payof a certain methodological imprecision and intentional ambiguity. Our work's perfectly contingent point of departure was the sudden perception of a resonance between the results of our research on Amazonian cosmopolitics-on its notion of a perspectivist multiplicity intrinsic to the real-and a well-known parable on the subject of the conquest of the Americans recounted by Lévi-Strauss in Race and History: In the Greater Antilles, some years after the discovery of America, while the Spaniards sent out investigating commissions to ascertain whether or not the natives had a soul, the latter were engaged in the drowning of white prisoners in order to verify, through prolonged watching, whether or not their corpses were subject to putrification. (L.-S. 1978[1952): 329) In this conflict between the two anthropologies, the author perceived a baroque allegory of the fact that one of the typical manifestations of human nature is the negation of its own generality. A kind of congenital avarice preventing the extension of the predicates of humanity to the species as a whole appears to be one of its predicates. In sum, ethnocentrism could be said to be like good sense, of which perhaps it is just the apperceptive moment: the best distributed thing in the world. The format of the lesson is familiar, but that does not lessen its sting. Overestimating one's own humanity to the detriment of the contemptible other's reveals one's deep resemblance with it. Since the other of the Same (of the European) shows itself to be the same as the Other's other (of the indigenous), the Same ends up unwittingly showing itself to be the same as the Other. The anecdote fascinated Lévi-Strauss enough for him to repeat it in Iristes Tropiques. But there he added a supplementary, ironic twist, this time noting a difference (rather than this re-semblance) between the parties. While the Europeans relied on the social sciences in their investigations of the humanity of the other, the Indians placed their faith in the natural sciences; and where the former proclaimed the Indians to be animals, the latter were content to suspect the others might be gods. "Both attitudes show equal ignorance," Lévi-Strauss concluded, "but the Indian's behavior certainly had greater dignity" (1992: 76). If this is really how things transpired,' it forces us to conclude that, despite being just as ignorant on the subject of the other, the other of the Other was not exactly the same as the other of the Same. We could even say that it was its exact opposite, if not for the fact that the relation between these two others of humanity-animality and divinity—is conceived in indigenous worlds in completely different terms than those we have inherited from Christianity. The rhetorical contrast Lévi-Strauss draws succeeds because it appeals to our cosmological hierarchies rather than those of the Taino. In any case, consideration of this disequilibrium was what led us to the hypothesis that Amerindian ontological regimes diverge from those widespread in the West precisely with regard to the inverse semiotic functions they respectively attribute to soul and body. The marked dimension for the Spanish was the soul, whereas the Indian emphasized the body. The Europeans never doubted that the Indians had bodies--animals have them too-and the Indians in turn never doubted that the Europeans had souls, since animals and the ghosts of the dead do as well. Thus the Europeans' ethnocentrism consisted in doubting that the body of the other contained a soul formally similar to the one inhabiting their own bodies, while the ethnocentrism of the Indians, on the contrary, entailed doubting that the others' souls or spirits could possess a body materially similar to theirs. • In the semiotic terms of Roy Wagner, a Melanesianist who will quickly reveal himself to be a crucial intercessor in the theory of Amerindian perspectivism, the body belongs to the innate or spontaneous dimension of European ontology ("nature"), which is the counter-invented result of an operation of conventionalist symbolization, while the soul would be the constructed dimension, the fruit of a "differentiating" symbol-ization that "specifies and renders concrete the conventional world by tracing radical distinctions and concretizing the singular individuals of this world" (Wagner 1981: 42). In indigenous worlds, on the contrary, the soul "is experienced as ... a manifestation of the conventional order implicit in everything" and "sums up the ways in which its possessor is similar to others, over and above the ways in which he differs from them" (Wagner 1981: 94); the body, on the contrary, belongs to the sphere of what comes from the responsibility of agents and is one of the fundamental figures of something that has to be constructed against a universal and innate ground of an "immanent humanity" (Wagner 1981: 86-9), In short, European praxis consists in "making souls" (and differentiating cultures) on the basis of a given corporeal-material ground nature while indigenous praxis consists in "making bodies" (and differentiating species) on the basis of a socio-spiritual continuum, itself also given ... but in myth, as we will see. Wagner's conceptually dense and quite original theoretical system resists didactic summary; thus we request that the reader directly engage its most elegant and realized presentation in The Invention of Culture. Grosso modo, the Wagnerian semiotic can be said to be a theory of human and nonhuman practice conceived as exhaustively consisting in the recipro-cal, recursive operation of two modes of symbolization: (1) a collectiv-izing, conventional (or literal) symbolism where signs are organized in standardized contexts (semantic domains, formal languages, etc.) to the extent that they are opposed to a heterogeneous plane of "referents". that is, they are seen as symbolizing something other than themselves; and (2) a differentiating, inventive (or figurative) mode in which the world of phenomena represented by conventional symbolization is understood to be constituted by "symbols representing themselves," that is, events that simultaneously manifest as symbols and referents, thereby dissolving the conventional contrast. It should be observed, first of all, that the world of referents or the "real" is defined here as a semiotic effect: what is other to a sign is another sign having the singular capacity of "representing itself." The mode of existence of actual entities qua events or occasions is a tautegory. It should be stressed that the contrast between the two modes is itself the result of a conventionalist operation (and perception): the distinction between invention and convention is itself conventional, but at the same time every convention is produced through a counter-invention. The contrast is thus intrinsically recursive, especially if we understand that human cultures are fundamentally in conflict over the mode of symbolization they (conventionally) privilege as an element appropriated for action or invention, in reserving to the other the function of the "given." Cultures, human macrosystems of conventions, are distinguished by what they define as belonging to the sphere of the responsibilities of agents-the mode of the constructed-and by what belongs (because it is counter-constructed as belonging) to the world of the given or non-constructed. The core of any and every set of cultural conventions is a simple distinction as to what kind of contexts- the nonconventionalized ones or those of convention itself- are to be deliberately articulated in the course of human action, and what kind of contexts are to be counter-invented as "motivation" under the conventional mask of "the given" or "the innate." Of course I...] there are only two possibilities: a people who deliberately differentiate as the form of their action will invariably counter-invent a motivating collectivity as "innate," and a people who deliberately collectivize will counter-invent a motivating differentiation in this way. (Wagner 1981: 51)
考点1. ”virtual affinity“必须译为”虚拟姻亲“或者”虚拟姻亲关系“,一种即可,禁止前后不一致 考点2.”transcendental empiricism“必须译为”先验经验论/主义”或“超验主义/经验论”,一种即可,禁止前后不一致 考点3.”Geophilosophy“必须译为”地理哲学“/“地理-哲学”/“地学哲学”,一种即可,禁止前后不一致 考点4. “baroque allegory”推荐译为“巴洛克(式)寓言” 考点5. “cannibal alterity”推荐译为“食人的他性”/“食人式他性”,一种即可,禁止前后不一致 考点6. multinaturalism 推荐译为 “多元自然论” 或 “多自然主义”,一种即可,禁止前后不一致 考点7. matrimonial alliance 推荐译为 “婚姻联盟” 或 “姻亲关系”,一种即可,禁止前后不一致
310de
学术论文
人文科学
106
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 葛兰西在论述意大利现代民族国家形成与发展中的政治领导权时,论及一个社会集团的霸权地位,体现在“统治”和“智识与道德”意义上的文化领导权:一个社会集团取得政权的首要条件就是之前已经在行使文化领导权,在其掌握政权之时,也必须延续以往的文化领导权。葛兰西明确地将“统治”和“意识形态霸权”(即文化领导权)作了区分,并强调了意识形态霸权对于取得政权和维系政权的重要意义,而意识形态还通过权力关系渗透到大众日常生活的现实意识,包括语言、常识与大众宗教,而这些意识形态要素和实践,都被整合到一个民俗体系中。至于统治阶级的意识形态霸权如何通过权力关系作用于大众的日常意识,葛兰西并没有系统地展开论述。英国马克思主义文化理论家威廉斯在葛兰西的基础上进一步论证了意识形态的作用机制,在这一机制中发挥决定性作用的不仅是观念和信仰的意识形态体系本身,还包括文化领导权所主导的社会生活的整体实践过程。而主导性的意识形态则可能遮蔽了“那一时期或社会中现实人们的那些相对复杂的、混合的、不完整的或未得到清晰表述的思想意识”。威廉斯的论述实际上肯定了葛兰西的意识形态霸权影响大众意识的观点,在此基础上又揭示了前者遮蔽后者的可能性,同时还在一定程度上避免了意识形态霸权所具有的意识形态决定论取向。至于大众意识的主体性,在意识形态霸权理论中则似乎付之阙如。 美国的中国学家姜士彬在研究明清时期的信息沟通方式时指出,葛兰西“意识形态霸权”所指涉的现代西方社会存在私人组织和国家权力机关的二元化,所以意识形态是通过社会中的报章、教会和学校来传播和灌输的。而明清时期的所有国家机关和民间组织都被士绅这同一个精英阶层控制,以至于统治阶级“霸权”达到了惊人的强度和范围,甚至此时的中国已经成为一个“学者政权”,他将其视为一个建立在特定意识形态基础上的文化整合的过程。由于姜士彬过度强调文化整合,所以没有在意识形态霸权如何遮蔽大众意识这一层面进一步拓展威廉斯的理论观点。 葛希芝、魏乐博在此基础上还借用了斯科特“弱者的日常抵抗”概念,论证了明清民众意识的日常抵抗形式和自主性,从而提出“民间意识形态”概念。斯科特继对东南亚农民反叛的研究之后发现了农民反抗的“日常”形式,例如磨洋工、装傻、逃避、假奉承、盗窃、散布谣言、放火、搞破坏等等,他称之为“弱者的武器”。他在关注反抗实践的同时,也关注村庄内的意识形态斗争,其核心的问题意识是讨论反抗和阶级斗争重要主题的意识形态支配问题。由于斯科特关注的是农民的日常反抗形式及其背后的意识形态问题,农民的这种大众意识形态主要是以日常反抗为基调的。而葛希芝、魏乐博及其合作者们所提出的民间意识形态概念除了包含斯科特所说的大众意识形态之外,还因明清时期士绅群体在民间社会的支配性影响力而将正统伦理观以小传统的形式纳人了民间意识形态。“民间意识形态”作为一个动态的社会进程,既呈现了姜士彬所说的文化整合机制,也凸显了民间抵抗的一面。这种民间抵抗大量地表现为一种日常反抗的形式,有时也会发展为一定规模的反叛,秘密教派可能会成为民众反叛的组织形式。当然,民众的日常抵抗和反叛也存在着意识形态与权力网络的内在一致性及特定的转化机制。韩书瑞的研究表明,白莲教不仅在教义方面与儒家伦理相契合,而且在经卷中常常提到老子和观音:在教派活动中,还可以提供与僧道、灵媒等宗教职业人员同样的服务。也正是这种正统化的民间宗教实践使得秘密教派具有实施广泛政治动员的意识形态基础。 本文的研究问题不在于探寻“民间意识形态”本身的思想史意义,而是追究“民间意识形态”的社会建构是如何呈现为历史本体论层面的“象征支配”的。格尔兹曾论证过,意识形态之于社会政治现实的意义就在于其高度象征性能将社会情势在其构架内进行解释并赋予其意义,还能使接受它的人产生自觉的认同感。格尔兹将意识形态视为一种文化体系,意识形态对社会行为的解释使其既具有高度的象征性,又是社会一体化的文化粘合剂。格尔兹作为阐释人类学大师,尤其强调意识形态象征的意义系统,而缺乏对意识形态作为文化体系实践机制的探讨。 布迪厄在实践理论中提出了“象征资本”的概念,以解释前资本主义经济的象征支配逻辑,其论证的核心就是经济资本如何转化为象征资本,但是这一转化过程奠定于物质基础之上,也由此产生了含情脉脉的道德面纱掩盖下的身份依附关系。其象征形式多表现为感激、效忠、尊敬、道义责任或良心债之类的道德规范和价值观,“是利用一个效忠群体所进行的物质和象征性投资,是一种象征支配”。这种发生在前资本主义经济中的象征支配的建立、维持或恢复必须运用乔装打扮、改头换面的“委婉化策略,这一策略的运用机制即“社会炼金术”的基本操作。“社会炼金术”呈现了既定秩序的合法化效应,亦即文化的再生产抑或物质再生产都行使着意识形态职能,使得依附和支配关系借助“委婉化”策略而处于隐蔽状态。在象征支配的运作逻辑中,经济基础与上层建筑的区分就显得过于简单和粗糙了。所以,不存在一个整齐划一的实体性意识形态,即使是葛兰西意义上的意识形态霸权,也是通过特定的媒介将权力关系渗透到民众日常意识和价值观中的,更不用说“民间意识形态”了。象征支配的运作逻辑中呈现出特定的分散性和整体性,一如“分散性宗教”在民众信仰中呈现为结构性与制度性的存在。 斯科特在评论葛兰西“意识形态霸权”的局限性时指出,应避免经济决定论的极端形式的弊病,赋予广义上的意识形态领域一定程度上的自主性,但同时也应避免走向意识形态决定论的另一极端。他沿着布迪厄象征支配的“委婉化”通过细致的田野工作探讨了马来西亚塞达卡村庄意识形态斗争的日常形式,主要表现即经济支配的“委婉化”,作为其后果之一的财产关系“委婉化”,总是有关象征的操纵、斗争与冲突的焦点。物质性面相的经济支配本身就是象征支配的过程,而象征支配的实践逻辑则蕴涵了意识形态的价值操控和权力关系的运作。 斯科特还就此提出了“双重象征操控”的概念,以进一步论证经济支配委婉化的运作逻辑也就是一种委婉化象征的互惠性的操纵。他以塞达卡的哈姆扎与其雇主哈吉·卡迪尔的关系作为案例分析了这一“双重象征操控”的机制,哈姆扎利用雇主将为其提供工作机会表述为“帮助”或“援助”,刻意以礼貌的语言“恳求”卡迪尔的“帮助”由此将自己的行动置于最有利的位置,以实现自己找工作或寻求贷款的目标,进而有意无意地使自己的行动策略为公共合法性的构建做出了贡献。“双重象征操控”强调了富人剥削的委婉化表述和“弱者”行动策略的“互惠”机制,富人或精英的支配需要“双重象征操控”的互惠机制才能得以实现。当然,这一“双重象征操控”的实现不是互惠原则所能概括的,斯科特关于支配与抵抗艺术的进一步研究表明,有权者极为关心如何使象征符合其支配形式,而从属者通常都具有充足的理由去帮助支配者维持其象征,或至少不会公开与之抵触,这就表现为支配、抵抗艺术中公开剧本和潜隐剧本之间复杂的权力关系。
葛兰西在论述意大利现代民族国家形成与发展中的政治领导权时,论及一个社会集团的霸权地位,体现在“统治”和“智识与道德”意义上的文化领导权:一个社会集团取得政权的首要条件就是之前已经在行使文化领导权,在其掌握政权之时,也必须延续以往的文化领导权。葛兰西明确地将“统治”和“意识形态霸权”(即文化领导权)作了区分,并强调了意识形态霸权对于取得政权和维系政权的重要意义,而意识形态还通过权力关系渗透到大众日常生活的现实意识,包括语言、常识与大众宗教,而这些意识形态要素和实践,都被整合到一个民俗体系中。至于统治阶级的意识形态霸权如何通过权力关系作用于大众的日常意识,葛兰西并没有系统地展开论述。英国马克思主义文化理论家威廉斯在葛兰西的基础上进一步论证了意识形态的作用机制,在这一机制中发挥决定性作用的不仅是观念和信仰的意识形态体系本身,还包括文化领导权所主导的社会生活的整体实践过程。而主导性的意识形态则可能遮蔽了“那一时期或社会中现实人们的那些相对复杂的、混合的、不完整的或未得到清晰表述的思想意识”。威廉斯的论述实际上肯定了葛兰西的意识形态霸权影响大众意识的观点,在此基础上又揭示了前者遮蔽后者的可能性,同时还在一定程度上避免了意识形态霸权所具有的意识形态决定论取向。至于大众意识的主体性,在意识形态霸权理论中则似乎付之阙如。 美国的中国学家姜士彬在研究明清时期的信息沟通方式时指出,葛兰西“意识形态霸权”所指涉的现代西方社会存在私人组织和国家权力机关的二元化,所以意识形态是通过社会中的报章、教会和学校来传播和灌输的。而明清时期的所有国家机关和民间组织都被士绅这同一个精英阶层控制,以至于统治阶级“霸权”达到了惊人的强度和范围,甚至此时的中国已经成为一个“学者政权”,他将其视为一个建立在特定意识形态基础上的文化整合的过程。由于姜士彬过度强调文化整合,所以没有在意识形态霸权如何遮蔽大众意识这一层面进一步拓展威廉斯的理论观点。 葛希芝、魏乐博在此基础上还借用了斯科特“弱者的日常抵抗”概念,论证了明清民众意识的日常抵抗形式和自主性,从而提出“民间意识形态”概念。斯科特继对东南亚农民反叛的研究之后发现了农民反抗的“日常”形式,例如磨洋工、装傻、逃避、假奉承、盗窃、散布谣言、放火、搞破坏等等,他称之为“弱者的武器”。他在关注反抗实践的同时,也关注村庄内的意识形态斗争,其核心的问题意识是讨论反抗和阶级斗争重要主题的意识形态支配问题。由于斯科特关注的是农民的日常反抗形式及其背后的意识形态问题,农民的这种大众意识形态主要是以日常反抗为基调的。而葛希芝、魏乐博及其合作者们所提出的民间意识形态概念除了包含斯科特所说的大众意识形态之外,还因明清时期士绅群体在民间社会的支配性影响力而将正统伦理观以小传统的形式纳人了民间意识形态。“民间意识形态”作为一个动态的社会进程,既呈现了姜士彬所说的文化整合机制,也凸显了民间抵抗的一面。这种民间抵抗大量地表现为一种日常反抗的形式,有时也会发展为一定规模的反叛,秘密教派可能会成为民众反叛的组织形式。当然,民众的日常抵抗和反叛也存在着意识形态与权力网络的内在一致性及特定的转化机制。韩书瑞的研究表明,白莲教不仅在教义方面与儒家伦理相契合,而且在经卷中常常提到老子和观音:在教派活动中,还可以提供与僧道、灵媒等宗教职业人员同样的服务。也正是这种正统化的民间宗教实践使得秘密教派具有实施广泛政治动员的意识形态基础。 本文的研究问题不在于探寻“民间意识形态”本身的思想史意义,而是追究“民间意识形态”的社会建构是如何呈现为历史本体论层面的“象征支配”的。格尔兹曾论证过,意识形态之于社会政治现实的意义就在于其高度象征性能将社会情势在其构架内进行解释并赋予其意义,还能使接受它的人产生自觉的认同感。格尔兹将意识形态视为一种文化体系,意识形态对社会行为的解释使其既具有高度的象征性,又是社会一体化的文化粘合剂。格尔兹作为阐释人类学大师,尤其强调意识形态象征的意义系统,而缺乏对意识形态作为文化体系实践机制的探讨。 布迪厄在实践理论中提出了“象征资本”的概念,以解释前资本主义经济的象征支配逻辑,其论证的核心就是经济资本如何转化为象征资本,但是这一转化过程奠定于物质基础之上,也由此产生了含情脉脉的道德面纱掩盖下的身份依附关系。其象征形式多表现为感激、效忠、尊敬、道义责任或良心债之类的道德规范和价值观,“是利用一个效忠群体所进行的物质和象征性投资,是一种象征支配”。这种发生在前资本主义经济中的象征支配的建立、维持或恢复必须运用乔装打扮、改头换面的“委婉化策略,这一策略的运用机制即“社会炼金术”的基本操作。“社会炼金术”呈现了既定秩序的合法化效应,亦即文化的再生产抑或物质再生产都行使着意识形态职能,使得依附和支配关系借助“委婉化”策略而处于隐蔽状态。在象征支配的运作逻辑中,经济基础与上层建筑的区分就显得过于简单和粗糙了。所以,不存在一个整齐划一的实体性意识形态,即使是葛兰西意义上的意识形态霸权,也是通过特定的媒介将权力关系渗透到民众日常意识和价值观中的,更不用说“民间意识形态”了。象征支配的运作逻辑中呈现出特定的分散性和整体性,一如“分散性宗教”在民众信仰中呈现为结构性与制度性的存在。 斯科特在评论葛兰西“意识形态霸权”的局限性时指出,应避免经济决定论的极端形式的弊病,赋予广义上的意识形态领域一定程度上的自主性,但同时也应避免走向意识形态决定论的另一极端。他沿着布迪厄象征支配的“委婉化”通过细致的田野工作探讨了马来西亚塞达卡村庄意识形态斗争的日常形式,主要表现即经济支配的“委婉化”,作为其后果之一的财产关系“委婉化”,总是有关象征的操纵、斗争与冲突的焦点。物质性面相的经济支配本身就是象征支配的过程,而象征支配的实践逻辑则蕴涵了意识形态的价值操控和权力关系的运作。 斯科特还就此提出了“双重象征操控”的概念,以进一步论证经济支配委婉化的运作逻辑也就是一种委婉化象征的互惠性的操纵。他以塞达卡的哈姆扎与其雇主哈吉·卡迪尔的关系作为案例分析了这一“双重象征操控”的机制,哈姆扎利用雇主将为其提供工作机会表述为“帮助”或“援助”,刻意以礼貌的语言“恳求”卡迪尔的“帮助”由此将自己的行动置于最有利的位置,以实现自己找工作或寻求贷款的目标,进而有意无意地使自己的行动策略为公共合法性的构建做出了贡献。“双重象征操控”强调了富人剥削的委婉化表述和“弱者”行动策略的“互惠”机制,富人或精英的支配需要“双重象征操控”的互惠机制才能得以实现。当然,这一“双重象征操控”的实现不是互惠原则所能概括的,斯科特关于支配与抵抗艺术的进一步研究表明,有权者极为关心如何使象征符合其支配形式,而从属者通常都具有充足的理由去帮助支配者维持其象征,或至少不会公开与之抵触,这就表现为支配、抵抗艺术中公开剧本和潜隐剧本之间复杂的权力关系。
考点1. “民间意识形态”推荐译为“Popular Ideology” 考点2. “弱者的日常抵抗”推荐译为“Everyday Forms of Peasant Resistance” 考点3. “问题意识”推荐译为“Problematic; Core Problematic” 考点4. “磨洋工”推荐译为“Foot-dragging; Slacking off” 考点5. “含情脉脉”推荐译为“Tender; Sentimental” 考点6. “付之阙如”推荐译为“(is) conspicuously absent; (appears to be) a lacuna” 考点7. “姜士彬” 必须译为“David Johnson” 考点8. “葛希芝” 必须译为“Hill Gates” 考点9. “统治”推荐译为“rule” 考点10. “霸权”推荐译为“hegemonic position” 考点11. “正统化的民间宗教实践”推荐译为“orthodox-style popular religious practice” 考点12. “隐蔽”推荐译为““to conceal; to hide”” 考点13. “民俗”推荐译为“folkloric system; a system of folk customs” 考点14. “整体性”推荐译为“wholeness; totality”
31b76
学术论文
社会科学
122
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: 1.1 Rationale of the study We live in a world of advertising. As potential consumers, we are endlessly bombarded with all kinds of product or service information from various media including newspapers, magazines, television, radio, posters and Internet, etc. Advertising provides a valuable service to society and its members, because it defines for consumers the meaning and the role of products, services, and institutions. It indicates the difference that exists between brands of products and alternative services, as well as the distinguishing characteristics of companies and institutions. Advertising also tells the consumer what a specific product, brand or service should do when it is used and thus helps him or her to understand and evaluate experience with the products and services that he or she uses. On the other hand, by making people aware of products, service and ideas, advertising promotes sales and profits. Finally, advertising is one of the major forces that are helping improve the standard of living around the world. Combined with all these communicational, marketing and social functions. Advertising becomes indispensable in the modern world. Naturally, advertisements in English have become an important means of communicating ideas, demonstrating a variety of linguistic features of its own. The present study attempts to examine these features at the lexical, syntactic and discourse levels, in the hope of bringing them to light and, thereby, offering help to advertisement writers and language learners. 1.2 Definition of advertising According to the Definition Committee of American Marketing Association(方薇, 1997:2), advertising is defined as follows: Advertising is the nonpersonal communication of information usually paid for and usually persuasive in nature about products, services or ideas by identified sponsors through the various media. 1.3 Focus of the present study Usually, advertising communicates information in three types: audio, visual, and language. It is a more common case that an advertisement is a mixture of the three. In radio advertisements, music is always accompanied by language; on TV and motion pictures, music and language illustration are mixed with each other. In magazines and newspapers, advertisements are a combination of pictures and language of written information. Although music and pictures can provide some hints, or create a kind of atmosphere, the information about the product is limited. Even worse, it may lead to misunderstanding. Thus, we may say that language in a way provides more exact, detailed and dependable information whereas music and pictures only act as a supplementary means in advertising. Advertising language, playing a role of communication and persuasion, has developed its own features. This paper will focus on the language features of English advertisements at lexical, syntactic and discourse levels. It is hoped, by a contrastive study of advertisements on three types of products (daily consumer goods, technical equipment and service), similarities and differences of the three types of advertisements will be summarized and possible reasons will be given in the light of the meaning, and function of language. 1.4 Sources of data All the advertisements studied in this paper are taken from English magazines. They are chosen from Time, People and Newsweek (issues from 1999-2000), because these three magazines have a huge circulation, covering all kinds of audience. Almost all kinds of advertisements can be found in these magazines. In order to get valuable information for the study, a corpus of 60 advertisements was built, which consists of 20 daily consumer goods ads, 20 technical equipment ads, and 20 service ads. Conclusions will be drawn through quantitative and qualitative studies of the data. 2 Lexical Features 2.1 Classification of advertising and its audience Generally speaking, advertisements can be divided into two types: public relation ads and commercial ads. The former tries to advocate reputation for a social group, whose purpose is to leave a favorable impression upon the potential audience. The latter leads to the act of purchasing the products or using the recommended service. Commercial ads are much more presented through mass media for the reason that manufacturers and companies are willing to spend a large sum of money to make a certain product known or to boost the image of a certain brand. In some cases, competitors, like Coca-cola and Perpsi, even spare no expense to launch advertising campaigns to win over the market share. Commercial advertising can also be divided according to the target audience into two groups: consumer advertising and business advertising. Most of the ads in the mass media are consumer advertisements. They are typically directed at consumers. By contrast, business advertising tends to be concentrated in specialized business publications, professional journals, trade shows targeting at a certain group of people involved in some business. Since consumer advertising is most accessible to common people, the present study on will focus on consumer advertising. The classification of advertising is clearly shown in the following graph: The bold parts show the scope of advertisements we study. Daily consumer goods are necessities of daily life, such as food, detergent, hygiene, etc. Technical equipment is technical toys and electric equipment such as camera, vehicle, hi-fi, etc. Service covers bank, insurance, fund, etc. Actually, advertising works effectively some of the time and doesn’t work other times. The single crucial reason that advertising does not work is that in specific instances the information it conveys never reaches the consumer at all, or is judged by the consumer to be either redundant, meaningless, or irrelevant. For example, a motorbike advertisement will probably be invisible to housewives on the lookout for new cutlery. Social status and individual interest decide that consumer goods ads are mainly targeting at women while technical equipment ads are largely aiming at men. The amount of shared knowledge between the advertiser and the audience together with the thinking habit of the audience directly influences the advertising language. Since products and audience change in every advertisement in order to achieve high advertising effectiveness, language used differs in different types of advertisements. Thus, in this paper we discuss not only the similarities of language shared by all types of advertisements but also differences of language used in different kinds of advertisements. 2.2 Similarities at the lexical level In order to make the information accessible to audience effectively, the choice of words in advertising is very cautious and skillful. The aim of the advertiser is quite specific. He wishes to capture the attention of the members of a mass audience and by means of impressive words to persuade them to buy a product or behave in a particular way, such as going to Hawaii for all their holiday needs. Both linguistic and psychological aspects are taken into consideration in the choice of words. Sharing the same purpose of advertising-to familiarize or remind consumers of the benefits of particular products in the hope of increasing sales, the techniques used at the lexical level by advertisers do not vary markedly. The following points are some prominent similarities.
1.1 Rationale of the study We live in a world of advertising. As potential consumers, we are endlessly bombarded with all kinds of product or service information from various media including newspapers, magazines, television, radio, posters and Internet, etc. Advertising provides a valuable service to society and its members, because it defines for consumers the meaning and the role of products, services, and institutions. It indicates the difference that exists between brands of products and alternative services, as well as the distinguishing characteristics of companies and institutions. Advertising also tells the consumer what a specific product, brand or service should do when it is used and thus helps him or her to understand and evaluate experience with the products and services that he or she uses. On the other hand, by making people aware of products, service and ideas, advertising promotes sales and profits. Finally, advertising is one of the major forces that are helping improve the standard of living around the world. Combined with all these communicational, marketing and social functions. Advertising becomes indispensable in the modern world. Naturally, advertisements in English have become an important means of communicating ideas, demonstrating a variety of linguistic features of its own. The present study attempts to examine these features at the lexical, syntactic and discourse levels, in the hope of bringing them to light and, thereby, offering help to advertisement writers and language learners. 1.2 Definition of advertising According to the Definition Committee of American Marketing Association(方薇, 1997:2), advertising is defined as follows: Advertising is the nonpersonal communication of information usually paid for and usually persuasive in nature about products, services or ideas by identified sponsors through the various media. 1.3 Focus of the present study Usually, advertising communicates information in three types: audio, visual, and language. It is a more common case that an advertisement is a mixture of the three. In radio advertisements, music is always accompanied by language; on TV and motion pictures, music and language illustration are mixed with each other. In magazines and newspapers, advertisements are a combination of pictures and language of written information. Although music and pictures can provide some hints, or create a kind of atmosphere, the information about the product is limited. Even worse, it may lead to misunderstanding. Thus, we may say that language in a way provides more exact, detailed and dependable information whereas music and pictures only act as a supplementary means in advertising. Advertising language, playing a role of communication and persuasion, has developed its own features. This paper will focus on the language features of English advertisements at lexical, syntactic and discourse levels. It is hoped, by a contrastive study of advertisements on three types of products (daily consumer goods, technical equipment and service), similarities and differences of the three types of advertisements will be summarized and possible reasons will be given in the light of the meaning, and function of language. 1.4 Sources of data All the advertisements studied in this paper are taken from English magazines. They are chosen from Time, People and Newsweek (issues from 1999-2000), because these three magazines have a huge circulation, covering all kinds of audience. Almost all kinds of advertisements can be found in these magazines. In order to get valuable information for the study, a corpus of 60 advertisements was built, which consists of 20 daily consumer goods ads, 20 technical equipment ads, and 20 service ads. Conclusions will be drawn through quantitative and qualitative studies of the data. 2 Lexical Features 2.1 Classification of advertising and its audience Generally speaking, advertisements can be divided into two types: public relation ads and commercial ads. The former tries to advocate reputation for a social group, whose purpose is to leave a favorable impression upon the potential audience. The latter leads to the act of purchasing the products or using the recommended service. Commercial ads are much more presented through mass media for the reason that manufacturers and companies are willing to spend a large sum of money to make a certain product known or to boost the image of a certain brand. In some cases, competitors, like Coca-cola and Perpsi, even spare no expense to launch advertising campaigns to win over the market share. Commercial advertising can also be divided according to the target audience into two groups: consumer advertising and business advertising. Most of the ads in the mass media are consumer advertisements. They are typically directed at consumers. By contrast, business advertising tends to be concentrated in specialized business publications, professional journals, trade shows targeting at a certain group of people involved in some business. Since consumer advertising is most accessible to common people, the present study on will focus on consumer advertising. The classification of advertising is clearly shown in the following graph: The bold parts show the scope of advertisements we study. Daily consumer goods are necessities of daily life, such as food, detergent, hygiene, etc. Technical equipment is technical toys and electric equipment such as camera, vehicle, hi-fi, etc. Service covers bank, insurance, fund, etc. Actually, advertising works effectively some of the time and doesn’t work other times. The single crucial reason that advertising does not work is that in specific instances the information it conveys never reaches the consumer at all, or is judged by the consumer to be either redundant, meaningless, or irrelevant. For example, a motorbike advertisement will probably be invisible to housewives on the lookout for new cutlery. Social status and individual interest decide that consumer goods ads are mainly targeting at women while technical equipment ads are largely aiming at men. The amount of shared knowledge between the advertiser and the audience together with the thinking habit of the audience directly influences the advertising language. Since products and audience change in every advertisement in order to achieve high advertising effectiveness, language used differs in different types of advertisements. Thus, in this paper we discuss not only the similarities of language shared by all types of advertisements but also differences of language used in different kinds of advertisements. 2.2 Similarities at the lexical level In order to make the information accessible to audience effectively, the choice of words in advertising is very cautious and skillful. The aim of the advertiser is quite specific. He wishes to capture the attention of the members of a mass audience and by means of impressive words to persuade them to buy a product or behave in a particular way, such as going to Hawaii for all their holiday needs. Both linguistic and psychological aspects are taken into consideration in the choice of words. Sharing the same purpose of advertising-to familiarize or remind consumers of the benefits of particular products in the hope of increasing sales, the techniques used at the lexical level by advertisers do not vary markedly. The following points are some prominent similarities.
考点1:“bringing them to light”推荐译为“揭示这些特点” 考点2:“identified sponsors”中的“sponsor”应译为“广告主”,因此完整译文应该是“署名广告主” 考点3:“on the lookout for”推荐译为“留意寻找” 考点4:“circulation”应译为“发行量” 考点5:“public relation ads”是广告分类的一个标准术语,应译为“公关广告”
32406
学术论文
人文科学
47
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Taking advantage of online data collection and game-based assessment, we characterized comprehensive developmental trajectories of spatial cognition and its malleability in the present study. We found an a synchrony between the development of spatial cognitive performance and that of its malleability, as the learning rate peaked about 10 years earlier than performance on mental rotation. On the other hand, malleability also depended on the maturation of spatial ability, as children under the age of 10 showed particularly low malleability, which likely resulted from their underdeveloped ability of mirror image discrimination. In short, by combining both coarse but massive online data and fine tuned experimental data, our study illustrates the interaction of spatial cognition and its malleability during development, which also sheds light on translational implications for STEM related educational practice. The developmental trajectory of mental rotation performance observed here is consistent with previous reports on development in adulthood of spatial abilities based on laboratory studies (Wilson et al., 1975). Also, the widespread malleability of mental rotation across the life span is in line with the reported malleability of spatial abilities in various developmental stages (Bavelier et al., 2012; Feng et al., 2007; Uttal, Miller, et al.,2013). The convergence with the findings of traditional laboratory studies (Germine et al., 2012; Hartshorne &Germine, 2015) endorses the reliability and validity of the online game based assessment approach in the present study. Importantly, our study extended previous studies in two aspects. First, our results filled the gap between the known developmental trends of spatial abilities in childhood (Frick et al., 2013; Hawes et al.,2015; Marmor, 1975) and adulthood (Wilson et al., 1975)by providing information on the transition between childhood and adulthood and by depicting a continuous development trajectory across the life span with the same task. Second, the present study comprehensively examined the malleability of spatial ability from childhood to adulthood, identifying late adolescence (16–20years old) as a critical period when individuals are most sensitive to training on mental rotation in particular and possibly spatial cognition in general. A particularly intriguing finding of the present study is the a synchrony between the development of performance and its malleability. The asynchronous developmental paces of malleability and performance have been documented in other cognitive domains, such as language acquisition (Snow & Hoefnagel Höhle, 1978),implying a general principle for the development of cognitive functions. For instance, the training effect from daily practice in spatial cognition may accumulate through development and thus lays the foundation for performance change. Supporting this idea, our results showed that in adolescent and adult age groups, individuals with higher performance also showed higher malleability. On the other hand, both the exploratory online test and the confirmatory experiment suggest that mirror image discrimination might be a gate keeping ability that constrained the malleability of young children. This result is consistent with previous findings that children at the age of 7 to 8 or even older cannot reliably perform above chance in mental rotation tasks involving mirror image discrimination (Aaron & Malatesha, 1974;Hawes et al., 2015). The particular difficulty of mirror image discrimination, which persists across the life span(Bornstein et al., 1978; Gibson et al., 1962; Gregory &McCloskey, 2010), suggests that it may involve unique cognitive mechanisms not implicated in the discrimination of rotated stimuli. Indeed, a previous study shows that children at the age of 6 show a steady decrease in error rate in discriminating 3D objects of different orientations, but the rate of mirror image confusion stays high at least until 7 years old (Hawes et al., 2015). Various cognitive mechanisms may contribute to the particular difficulty of mirror image discrimination of children. For instance, Paschke et al. (2012) suggest that mirrored images fit less to the top-down expectation; Hamm et al. (2004) propose that mirror image discrimination requires an additional “flip” of the stimuli out oft he picture plane, which requires more complex transformation of mental representation; McCloskey et al.(2006) suggest that mirror image discrimination requires additionally representing polarity correspondence between the object and the extrinsic axes, which likely poses challenges to young children (see also Gregory& McCloskey, 2010). Our study revealed that the ages between 8 and 9 were critical for developing the ability of mirror image discrimination. Therefore, future studies can test these three hypotheses with children at this age by examining whether mirror image discrimination co­exists with the contributing cognitive functions proposed by these hypotheses. Indeed, the present study suggests that mirror image discrimination might hinder mental rotation performance and its malleability in a wider range of the life span. Visual inspection of Figure4c indicates that the high performers across age groups passed the mirror image trials without much difficulty and performance improves continually, resulting in a more linear shaped learning curve compared with the low performers. In contrast, the low performers needed a substantial amount of training to succeed in mirror image trials; therefore, the learning curve was flattened as an asymptotic line at the point of the mirror image trials. The finding that high malleability of spatial cognition occurred in late adolescence is of potentially great translational implication because of the transferability of mental rotation training effects into other spatial skills(Meneghetti et al., 2016; Terlecki & Newcombe, 2005;Uttal, Meadow, et al., 2013) as well as into STEM related performance (Stransky et al., 2010). That is, training on spatial abilities may be most effective for adolescents and young adults, whereas elementary school children may not benefit as much from it. However, the training should be merit based because there were significant individual differences in malleability as revealed in this study. Finally, the training may not be fully effective until the gatekeeping ability is mastered. In addition, the present study replicated gender difference in mental rotation in the literature with a large sample. Although the male advantage in mental rotation has been well established in the literature, the source of this advantage has been a subject of debate, with arguments emphasizing either biological or experience/cultural factors (Miller & Halpern, 2014). The late emergence of gender difference observed in our study (i.e.,its presence in adolescence and early adulthood but not in childhood) implies the influence of experience related and/or environmental factors, consistent with findings of cross national research (e.g., Levine et al.,2005) and studies on gender stereotype threat (e.g.,Schmader, 2010). For instance, males might be encouraged more to participate in spatial related tasks since childhood than females, and sufficient practice leads to performance advantage in adolescence, which might encourage further participation in spatial related tasks and a persistent male advantage in the following developmental stages. However, it is also possible that the gradual unfolding of male advantage could be the result of late on set hormonal influence or sex dependent brain polymorphism, such as those taking place during and after puberty, though evidence of such influence is mixed in the literature (e.g., Herlitz et al., 2013). Further, the experience related and/or environmental factors may interact with biological factors (Eagly & Wood,2013), as gender dependent educational experience and cultural expectation might facilitate the expression of phenotypes prescribed by biological factors. In sum, by combining both a coarse online test and a finetuned experiment, we depicted comprehensive developmental trajectories of spatial cognition and its malleability, which invites research on the development of other cognitive domains. However, our results should be interpreted with caution given several limitations with online studies. First, the participants were self-selected, and the sample composition was certainly biased by that of the user population of the host app. Also, the present study did not verify the accuracy of age reports made by the participants because we were unable to devise an effective safeguard/check for largescale online game-based data collection without making participants uncomfortable about their privacy. Therefore, some age reports might be inaccurate. However, this was unlikely to have distorted the pattern of the reported developmental trajectory because, given the large sample size of the present study, the noise caused by potential inaccurate age reports would have affected our estimation of all the age groups unselectively. Admittedly, the potential presence of inaccurate age reports may have led to an underestimation of developmental difference between age groups, and follow up small scale laboratory-based experiments are necessary to verify findings from the online experiment. Second, in the game-based assessment, participants reported their age information by selecting one of the age bins with the resolution of 5 years old (except theunder10 age group). Though this is a common practice in online game applications, it admittedly prevented us from depicting the developmental trajectory with finer resolution. This is particularly regrettable for the under10 age group and the 11–15 age group, in which rapid cognitive development in spatial cognition takes place. To some extent, the pen and paper test provided the missing piece by allowing us to analyze mental rotation ability between the ages of 6 and 13 to depict the change in mirror image discrimination ability during this period. More dedicated investigations are desirable to depict the developmental trajectory at this age range with finer resolution. Third, the online game-based assessment allowed us no control over the time span between participations or between participation activities. This might have led to additional noise in estimating the malleability of each age group. However, the variation in time span was unlikely to affect the relation between behavioral performance and its malleability of each age group differently. Fourth, online studies provide only cross sectional data. Therefore, follow-up longitudinal studies on the critical developmental periods identified by online studies are also necessary to rule out confounding factors such as the cohort effect. Fifth, in the present study, the malleability of spatial cognition was measured by the learning rate among multiple plays on mental rotation; therefore, malleability at the behavioral level may not fully capture plasticity in the neural substrates of mental rotation that may additionally reflect contributions from many factors other than training, such as lesion and sociocultural context. Future neuroimaging studies on the development of neural plasticity will likely provide an interesting comparison with the malleability solely from behavioral training. Finally, the findings of an asynchrony between performance and malleability and that of the coincidence between mirror image discrimination failure and low malleability of spatial cognition did not present causal links, and future work is needed to examine the causality among these factors.
Taking advantage of online data collection and game-based assessment, we characterized comprehensive developmental trajectories of spatial cognition and its malleability in the present study. We found an a synchrony between the development of spatial cognitive performance and that of its malleability, as the learning rate peaked about 10 years earlier than performance on mental rotation. On the other hand, malleability also depended on the maturation of spatial ability, as children under the age of 10 showed particularly low malleability, which likely resulted from their underdeveloped ability of mirror image discrimination. In short, by combining both coarse but massive online data and fine tuned experimental data, our study illustrates the interaction of spatial cognition and its malleability during development, which also sheds light on translational implications for STEM related educational practice. The developmental trajectory of mental rotation performance observed here is consistent with previous reports on development in adulthood of spatial abilities based on laboratory studies (Wilson et al., 1975). Also, the widespread malleability of mental rotation across the life span is in line with the reported malleability of spatial abilities in various developmental stages (Bavelier et al., 2012; Feng et al., 2007; Uttal, Miller, et al.,2013). The convergence with the findings of traditional laboratory studies (Germine et al., 2012; Hartshorne &Germine, 2015) endorses the reliability and validity of the online game based assessment approach in the present study. Importantly, our study extended previous studies in two aspects. First, our results filled the gap between the known developmental trends of spatial abilities in childhood (Frick et al., 2013; Hawes et al.,2015; Marmor, 1975) and adulthood (Wilson et al., 1975)by providing information on the transition between childhood and adulthood and by depicting a continuous development trajectory across the life span with the same task. Second, the present study comprehensively examined the malleability of spatial ability from childhood to adulthood, identifying late adolescence (16–20years old) as a critical period when individuals are most sensitive to training on mental rotation in particular and possibly spatial cognition in general. A particularly intriguing finding of the present study is the a synchrony between the development of performance and its malleability. The asynchronous developmental paces of malleability and performance have been documented in other cognitive domains, such as language acquisition (Snow & Hoefnagel Höhle, 1978),implying a general principle for the development of cognitive functions. For instance, the training effect from daily practice in spatial cognition may accumulate through development and thus lays the foundation for performance change. Supporting this idea, our results showed that in adolescent and adult age groups, individuals with higher performance also showed higher malleability. On the other hand, both the exploratory online test and the confirmatory experiment suggest that mirror image discrimination might be a gate keeping ability that constrained the malleability of young children. This result is consistent with previous findings that children at the age of 7 to 8 or even older cannot reliably perform above chance in mental rotation tasks involving mirror image discrimination (Aaron & Malatesha, 1974;Hawes et al., 2015). The particular difficulty of mirror image discrimination, which persists across the life span(Bornstein et al., 1978; Gibson et al., 1962; Gregory &McCloskey, 2010), suggests that it may involve unique cognitive mechanisms not implicated in the discrimination of rotated stimuli. Indeed, a previous study shows that children at the age of 6 show a steady decrease in error rate in discriminating 3D objects of different orientations, but the rate of mirror image confusion stays high at least until 7 years old (Hawes et al., 2015). Various cognitive mechanisms may contribute to the particular difficulty of mirror image discrimination of children. For instance, Paschke et al. (2012) suggest that mirrored images fit less to the top-down expectation; Hamm et al. (2004) propose that mirror image discrimination requires an additional “flip” of the stimuli out oft he picture plane, which requires more complex transformation of mental representation; McCloskey et al.(2006) suggest that mirror image discrimination requires additionally representing polarity correspondence between the object and the extrinsic axes, which likely poses challenges to young children (see also Gregory& McCloskey, 2010). Our study revealed that the ages between 8 and 9 were critical for developing the ability of mirror image discrimination. Therefore, future studies can test these three hypotheses with children at this age by examining whether mirror image discrimination co­exists with the contributing cognitive functions proposed by these hypotheses. Indeed, the present study suggests that mirror image discrimination might hinder mental rotation performance and its malleability in a wider range of the life span. Visual inspection of Figure4c indicates that the high performers across age groups passed the mirror image trials without much difficulty and performance improves continually, resulting in a more linear shaped learning curve compared with the low performers. In contrast, the low performers needed a substantial amount of training to succeed in mirror image trials; therefore, the learning curve was flattened as an asymptotic line at the point of the mirror image trials. The finding that high malleability of spatial cognition occurred in late adolescence is of potentially great translational implication because of the transferability of mental rotation training effects into other spatial skills(Meneghetti et al., 2016; Terlecki & Newcombe, 2005;Uttal, Meadow, et al., 2013) as well as into STEM related performance (Stransky et al., 2010). That is, training on spatial abilities may be most effective for adolescents and young adults, whereas elementary school children may not benefit as much from it. However, the training should be merit based because there were significant individual differences in malleability as revealed in this study. Finally, the training may not be fully effective until the gatekeeping ability is mastered. In addition, the present study replicated gender difference in mental rotation in the literature with a large sample. Although the male advantage in mental rotation has been well established in the literature, the source of this advantage has been a subject of debate, with arguments emphasizing either biological or experience/cultural factors (Miller & Halpern, 2014). The late emergence of gender difference observed in our study (i.e.,its presence in adolescence and early adulthood but not in childhood) implies the influence of experience related and/or environmental factors, consistent with findings of cross national research (e.g., Levine et al.,2005) and studies on gender stereotype threat (e.g.,Schmader, 2010). For instance, males might be encouraged more to participate in spatial related tasks since childhood than females, and sufficient practice leads to performance advantage in adolescence, which might encourage further participation in spatial related tasks and a persistent male advantage in the following developmental stages. However, it is also possible that the gradual unfolding of male advantage could be the result of late on set hormonal influence or sex dependent brain polymorphism, such as those taking place during and after puberty, though evidence of such influence is mixed in the literature (e.g., Herlitz et al., 2013). Further, the experience related and/or environmental factors may interact with biological factors (Eagly & Wood,2013), as gender dependent educational experience and cultural expectation might facilitate the expression of phenotypes prescribed by biological factors. In sum, by combining both a coarse online test and a finetuned experiment, we depicted comprehensive developmental trajectories of spatial cognition and its malleability, which invites research on the development of other cognitive domains. However, our results should be interpreted with caution given several limitations with online studies. First, the participants were self-selected, and the sample composition was certainly biased by that of the user population of the host app. Also, the present study did not verify the accuracy of age reports made by the participants because we were unable to devise an effective safeguard/check for largescale online game-based data collection without making participants uncomfortable about their privacy. Therefore, some age reports might be inaccurate. However, this was unlikely to have distorted the pattern of the reported developmental trajectory because, given the large sample size of the present study, the noise caused by potential inaccurate age reports would have affected our estimation of all the age groups unselectively. Admittedly, the potential presence of inaccurate age reports may have led to an underestimation of developmental difference between age groups, and follow up small scale laboratory-based experiments are necessary to verify findings from the online experiment. Second, in the game-based assessment, participants reported their age information by selecting one of the age bins with the resolution of 5 years old (except theunder10 age group). Though this is a common practice in online game applications, it admittedly prevented us from depicting the developmental trajectory with finer resolution. This is particularly regrettable for the under10 age group and the 11–15 age group, in which rapid cognitive development in spatial cognition takes place. To some extent, the pen and paper test provided the missing piece by allowing us to analyze mental rotation ability between the ages of 6 and 13 to depict the change in mirror image discrimination ability during this period. More dedicated investigations are desirable to depict the developmental trajectory at this age range with finer resolution. Third, the online game-based assessment allowed us no control over the time span between participations or between participation activities. This might have led to additional noise in estimating the malleability of each age group. However, the variation in time span was unlikely to affect the relation between behavioral performance and its malleability of each age group differently. Fourth, online studies provide only cross sectional data. Therefore, follow-up longitudinal studies on the critical developmental periods identified by online studies are also necessary to rule out confounding factors such as the cohort effect. Fifth, in the present study, the malleability of spatial cognition was measured by the learning rate among multiple plays on mental rotation; therefore, malleability at the behavioral level may not fully capture plasticity in the neural substrates of mental rotation that may additionally reflect contributions from many factors other than training, such as lesion and sociocultural context. Future neuroimaging studies on the development of neural plasticity will likely provide an interesting comparison with the malleability solely from behavioral training. Finally, the findings of an asynchrony between performance and malleability and that of the coincidence between mirror image discrimination failure and low malleability of spatial cognition did not present causal links, and future work is needed to examine the causality among these factors.
考点1: “ mirror image trials ” 推荐译为 “ 镜像实验”在本文语境中,可以发现trials实际上指代的是实验. 考点2: “ game-based assessment ” 推荐译为 “ 基于游戏的测评 ” 考点3: “ noise” 推荐译为 “ 随机误差 ”需确保“随机性”限定词不能省略 考点4 : “ general principle ” 在本文语境中推荐译为 推荐译为 “ 一般规律 ” 考点5: “ the reported ” 在本文语境中推荐译为 “ 报告 ”
32ddd
学术论文
社会科学
105
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 在音乐综艺《歌手2025》第四期,新生代实力女歌手单依纯甩着红发,对李荣浩创作的歌曲《李白》进行了颠覆性改编。打破原曲乡村摇滚框架,融入电子音效、游戏术语、即兴念白等元素,“诗人李白”被重构为“游戏李白”,她的改编犹如一枚重磅炸弹,成为近期全网热议焦点。单依纯念出的“我本是辅助,今晚来打野”“如何呢?又能怎?”等新增歌词在B站刷屏,而此前林志炫同样在《歌手2025》中精心设计的戏曲版《悟空》,却只留下“鹦鹉侠”的调侃标签,没有激起水花。 同出一部音综,都是“魔改”歌曲,为何命运截然不同?流行音乐研究专业博士、杭州师范大学副教授赵朴在接受记者采访时,给出了犀利解析:成功的“魔改”必须同时具备有梗、有趣、有料三重基因。“‘魔改’需要让歌曲变得有一点儿面目全非,又不会完全认不出原作,如果力度不够大,就成平平无奇的翻唱。而真正出圈的改编,还必须同时满足互联网传播的网感、音乐设计的巧思、精神内核的重构。” 有梗:点燃破圈传播的网感引擎 在互联网传播时代,一首歌曲是否有梗,往往成为作品能否破圈的爆破点。单依纯版《李白》将游戏术语“打野”“辅助”植入歌词,节目播出后48小时内催生16个热搜话题。那反复吟唱的“如何呢?又能怎?”恰如当代青年面对焦虑的戏谑解药,不仅让办公室白领、校园学子自发玩梗,甚至吸引了从不看音综的“路人”侧目。“《李白》的新增歌词十分具有网感,通过纯享版、切片版给大众提供了社交谈资。”赵朴指出,相较之下,林志炫的《悟空》虽在唱腔、妆造上倾注心血,却因缺乏具象记忆点,最终沦为曲高和寡的艺术实验。 也有曾经破圈的“魔改”歌曲深谙此道。比如麻园诗人在《乐队的夏天3》根据王心凌同名歌曲改编的《彩虹的微笑》,以笑中带泪、苦乐参半的真挚表达,触动了很多听者。当主唱苦果用撕裂般的哭腔唱甜歌、台上的小男孩翻起跟头,这童年与现实的残酷对撞,瞬间引发很多观众包括当时台下歌手大张伟的泪点。“我们面对生活的痛苦所做出的种种努力,无非只是想在这音乐里做一个翻跟头的孩子而已,因为那一刻叫作自由。” 有趣:提供新鲜奇妙的听觉体验 一首“魔改”歌曲如何让听众在赏乐过程中,感受到趣味性?这需要整体编曲与演唱设计的巧思。比如《李白》中后段插入的失真金属和弦,如同游戏特效般刺破耳膜;而单依纯刻意保留的破音、真假声转换时的瑕疵,恰似“酒中仙”李白“醉后失态”的声音化呈现。这种介于控制和失控之间的破绽,反而让技术缺陷转化为艺术表达的情感载体,使很多听众在“难评”的猎奇感中反复品味。 互联网知名乐评人耳帝直言:“在《歌手2025》大量无趣的选歌与保守改编的局面下,单依纯的尝试很值得鼓励,因为这样节目才会好看。在敢玩的同时,她依然未失去唱功的展示,最后还有一个高质量的强声E5。”但趣味也需要分寸感。比如有网友指出《李白》的缺陷:“歌曲当中用了大量游戏术语,作为非游戏玩家,听得一头雾水。”相较而言,二手玫瑰在改编《偶遇》时,通过巴乌、木鱼等冷门乐器构建异质音场,营造了静谧的宛如佛乐的氛围,既保留了音乐的新鲜感,又不设理解门槛,因此得到了更多专业乐评人的认可。 有料:构建具有深度的精神内核 什么样的“魔改”能够超越暂时的网红效应,成为值得细品的文化现象?必然得有对“料”的坚守----将原本的歌曲挖掘出更深的内涵,或者构建出新的意义。 比如B站上的音乐博主唐宋摇滚,穿着绯色大袖襕袍、头戴展脚幞头,抱一把木吉他,操着烟嗓唱歌,新近翻唱伍佰的《last dance(最后一舞)》播放量已接近450万。在旋律朗朗上口的基础上,唐宋摇滚的新写歌词被粉丝誉为“百万填词”。战场上的金戈铁马、建功立业的豪情壮志、儿女情仇的优柔慨叹,都寓于激情饱满的唱段之间。“没点文化,还真听不懂他玩的梗。”网友感叹,跟着歌声摇摆,得以一窥大宋风云。
在音乐综艺《歌手2025》第四期,新生代实力女歌手单依纯甩着红发,对李荣浩创作的歌曲《李白》进行了颠覆性改编。打破原曲乡村摇滚框架,融入电子音效、游戏术语、即兴念白等元素,“诗人李白”被重构为“游戏李白”,她的改编犹如一枚重磅炸弹,成为近期全网热议焦点。单依纯念出的“我本是辅助,今晚来打野”“如何呢?又能怎?”等新增歌词在B站刷屏,而此前林志炫同样在《歌手2025》中精心设计的戏曲版《悟空》,却只留下“鹦鹉侠”的调侃标签,没有激起水花。 同出一部音综,都是“魔改”歌曲,为何命运截然不同?流行音乐研究专业博士、杭州师范大学副教授赵朴在接受记者采访时,给出了犀利解析:成功的“魔改”必须同时具备有梗、有趣、有料三重基因。“‘魔改’需要让歌曲变得有一点儿面目全非,又不会完全认不出原作,如果力度不够大,就成平平无奇的翻唱。而真正出圈的改编,还必须同时满足互联网传播的网感、音乐设计的巧思、精神内核的重构。” 有梗:点燃破圈传播的网感引擎 在互联网传播时代,一首歌曲是否有梗,往往成为作品能否破圈的爆破点。单依纯版《李白》将游戏术语“打野”“辅助”植入歌词,节目播出后48小时内催生16个热搜话题。那反复吟唱的“如何呢?又能怎?”恰如当代青年面对焦虑的戏谑解药,不仅让办公室白领、校园学子自发玩梗,甚至吸引了从不看音综的“路人”侧目。“《李白》的新增歌词十分具有网感,通过纯享版、切片版给大众提供了社交谈资。”赵朴指出,相较之下,林志炫的《悟空》虽在唱腔、妆造上倾注心血,却因缺乏具象记忆点,最终沦为曲高和寡的艺术实验。 也有曾经破圈的“魔改”歌曲深谙此道。比如麻园诗人在《乐队的夏天3》根据王心凌同名歌曲改编的《彩虹的微笑》,以笑中带泪、苦乐参半的真挚表达,触动了很多听者。当主唱苦果用撕裂般的哭腔唱甜歌、台上的小男孩翻起跟头,这童年与现实的残酷对撞,瞬间引发很多观众包括当时台下歌手大张伟的泪点。“我们面对生活的痛苦所做出的种种努力,无非只是想在这音乐里做一个翻跟头的孩子而已,因为那一刻叫作自由。” 有趣:提供新鲜奇妙的听觉体验 一首“魔改”歌曲如何让听众在赏乐过程中,感受到趣味性?这需要整体编曲与演唱设计的巧思。比如《李白》中后段插入的失真金属和弦,如同游戏特效般刺破耳膜;而单依纯刻意保留的破音、真假声转换时的瑕疵,恰似“酒中仙”李白“醉后失态”的声音化呈现。这种介于控制和失控之间的破绽,反而让技术缺陷转化为艺术表达的情感载体,使很多听众在“难评”的猎奇感中反复品味。 互联网知名乐评人耳帝直言:“在《歌手2025》大量无趣的选歌与保守改编的局面下,单依纯的尝试很值得鼓励,因为这样节目才会好看。在敢玩的同时,她依然未失去唱功的展示,最后还有一个高质量的强声E5。”但趣味也需要分寸感。比如有网友指出《李白》的缺陷:“歌曲当中用了大量游戏术语,作为非游戏玩家,听得一头雾水。”相较而言,二手玫瑰在改编《偶遇》时,通过巴乌、木鱼等冷门乐器构建异质音场,营造了静谧的宛如佛乐的氛围,既保留了音乐的新鲜感,又不设理解门槛,因此得到了更多专业乐评人的认可。 有料:构建具有深度的精神内核 什么样的“魔改”能够超越暂时的网红效应,成为值得细品的文化现象?必然得有对“料”的坚守----将原本的歌曲挖掘出更深的内涵,或者构建出新的意义。 比如B站上的音乐博主唐宋摇滚,穿着绯色大袖襕袍、头戴展脚幞头,抱一把木吉他,操着烟嗓唱歌,新近翻唱伍佰的《last dance(最后一舞)》播放量已接近450万。在旋律朗朗上口的基础上,唐宋摇滚的新写歌词被粉丝誉为“百万填词”。战场上的金戈铁马、建功立业的豪情壮志、儿女情仇的优柔慨叹,都寓于激情饱满的唱段之间。“没点文化,还真听不懂他玩的梗。”网友感叹,跟着歌声摇摆,得以一窥大宋风云。
考点1:“鹦鹉侠”应译为“Parrot Man”。 考点2:“激起水花”应译为“ barely made a ripple”。 考点3:“魔改”应译为“drastic modification”。 考点4:“有料”应译为“rich-contented”。 考点5:“出圈”应译为“go viral”。 考点6:“网感”应译为“internet sensibility”。 考点7:“戏谑解药”应译为“a playful antidote”。 考点8:“曲高和寡”应译为“too highbrow to be popular”。 考点9:“麻园诗人”应译为“Mayuan Poet”。 考点10:“乐队的夏天3”应译为“The Big Band 3”。 考点11:“难评”应译为“hard to describe”。 考点12:“展脚幞头”应译为“zhanjiao futou”。 考点13:“襕袍”应译为“lanpao”。 考点14:“大宋风云”应译为“the epic sagas of the great Song Dynasty”。
32ed1
新闻资讯
新闻报道
60
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: The paper begins with a discussion of feasting in small-scale societies and suggests that recognizing the manner of articulation of three particular variables-the scale of participation and financing, the frequency and structure of occurrence, and the food resources used-is an effective starting point for gauging how politically charged feasts may have been, the actual social role feasting may have played, and the variation in feasting practices across space and through time. I then present the evidence for feasting in the Southwest from A.D. 850 to present, focusing particular attention on these three variables. I conclude the paper with a case study that highlights feasting as a political resource that is linked with, but subordinate to, other resources of power, especially the control of rituals and of the knowledge necessary to perform them. There are two sides to all communal behaviors: one that integrates and one that differentiates. Communal feasts are no exception. By their very nature, feasts bring people together to experience one of life's biological necessities in a communal, social manner. The communal exchange and sharing of food may be considered one of the most fundamental human transactions that, through social interaction and exchange, promotes social integration. This integrative aspect of feasting can have significant social and economic consequences. As Rappaport and Ford have noted, feasting can be an important mechanism for redistributing food among community members, and thus can be instrumental in promoting economic and social interdependence among community members. At the same time, to varying degrees, feasting can be an active context in which social hierarchy is established. Indeed, the hosting of feasts may be a particularly effective means through which to demonstrate one's economic and political abilities, and to engender prestige and the support of followers. Numerous aspects of feasting operate as public counting and ordering devices, which in turn reduce the vagueness of a social and political situation by promoting social comparison. For example, depending upon the quantity and quality of resources mobilized for communal feasts, and the frequency with which they are mobilized, feasting can be a quantitative measure of the abilities of the host as an efficient, skillful, vital, and generous leader. Since organizing and financing communal feasts often involves the cooperation of multiple individuals (e.g., communal hunts or large-scale domestic animal slaughters often precede communal feasts), large and elaborate feasts indicate that the host has control over the labor of others and is an efficient mobilizer of cooperative effort and, thus, a potential leader. Moreover, the degree to which the host successfully coordinates intercommunity participation at feasts is a particularly strong incentive for community allegiance and support, especially in situations involving supra-community alliances, exchange and marriage relations, conflict, and competition notes, "when food distributions are financed from outside,such as in many Highland New Guinea Big-man societies, abundance of food also is an indicator of a broad network of supportive social ties." Finally, if rare, exotic, or highly valued resources are presented and consumed in a feast, both the host's success as a procurer of these resources and, potentially, the host's control of sacred knowledge (i.e., the knowledge of how to acquire the exotic or rare resources) may be demonstrated by hosting a particularly successful feast.These display/performance aspects of communal feasting that promote social comparison operate in tandem with feasting's more fundamental capacity as a vehicle for engendering political support by creating debt-obligations through the unequal imparting of gifts and food). The creation of debt obligation is, in many anthropologists' eyes, a fundamental aspect of feasting that makes it a potent political vehicle . However, indebtedness may not be an effective political resource unless there are social or technological mechanisms in place that allow for the monopolization of the resources necessary to finance a feast so that a number of potential followers incur debt without the means for repayment by either paying off the debt or financing their own feast. It is only under these conditions that feasts may be used effectively to amplify and maintain social differences over the long term (e.g., a generation), regardless of the productivity of the environment. In addition, financing a feast may place the host in considerable debt to the host's supporters, a situation that may actually lessen the status of the host as an independent leader. Communal feasting also may be part of healing ceremonies, which are not usually structured by ritual cycles but which must occur on a fairly regular basis to prevent or cure illness. The Blessingway ceremony of the Navajo, for example, involves the feeding of the audience, which can sometimes number in the thousandsThese points are addressed below.There is tremendous variation in what has been termed feasting in middle-range societies. Most discussions of this variation have been couched in terms of ideal "types".The approach taken here, while not denying the utility of types as analytical constructs, focuses more on three particular dimensions of feasting that can combine in a number of ways to produce a variety of social and economic effects: 1) the scale of participation and financing; 2) the frequency and structure of occurrence; and 3) the food resources used. Communal feasts can range in scale from those in which a single household sponsors a feast for bilateral relatives or other households to those hosted by entire villages and offered to members of multiple communities. The scale of most communal feasts ranks somewhere between these two extremes. Among the Mae Enga and Chimbu of Highland New Guinea, feasts may be hosted by tribes (made up of clans), clans, subclans, or lineages. Many feasts in Puebloan communities are sponsored by multiple households within a community, with much of the community invited (or obligated) to participate . To a large extent, the means through which a feast is financed, and ultimately the purpose of the feast, determines its scale. Here it is useful to distinguish between household-level finance and supra-household finance . Assuming there are inherent limits on home production as a means of expanding the resource base of a feast, the establishment of social relations outside the lineage or household through sodality membership, marriage, exchange, or indebtedness all become paramount to hosting and financing largescale feasts. Postcontact Mae Enga big men of highland New Guinea, for example, may draw upon upwards of 2,000 followers for a single feast . The greater the scale of the feast, the greater the potential for prestige enhancement on the part of sponsors. Moreover, using feasting as a recurrent social practice to enhance prestige usually requires financing on a larger scale than that of the single household or lineage (Young 1985). Scale also may relate to whether feasts are primarily "intra-communal" or "inter-communal" affairs (although this distinction is often hard to make ethnographically.) Hayden (1995), for instance, delineates within-group feasts, which he terms "solidarity feasts," and between-group feasts, which are either "reciprocal" or "competitive" feasts. Reciprocal feasts function primarily to create alliances and promote cooperation among groups, while competitive feasts are hosted primarily to enhance the prestige, and political and economic support, of an aggrandizer and his/her followers or to negotiate and reaffirm extra-community political ties and agreement.Many within-group feasts have the principal effect of facilitating social integration or redistributing food among households within the village . Feasts of this type are usually financed by multiple households (often anonymously) and tend to be relatively non-competitive in nature. Compared to situations in which sponsors compete to enhance their prestige by contributing more to a feast than other sponsors, the inherent anonymity of, and lack of direct competition for, prestige in many intra-community integrative feasts tends to limit the quanitity of resources mobilized. In contrast, when feasts involve more than one community, cooperative effort on the part of village coresidents often facilitates the financing of larger and more elaborate feasts than is possible in intra-village contexts. It is in the context of inter-community (or large-scale) feasts that the prestige of aggrandizers becomes greatly enhanced relative to their followers.Much of the variation in occurrence of feasts is related to the degree to which feasting is regulated by the structure of ritual. Feasting often is associated with calendrical rituals that occur at prescribed times of the year. The kyanakwe feasting ceremony of the Zuni, for example, is held four times a year. Many feasts are held only once every several years. The Kaiko ritual of the Tsembaga of New Guinea may take place only once every 7 to 15 years . The scale of these feasts is smaller than that of unregulated or ad hoc feasts. Kaiko ritual feasts involve the participation of hundreds of people from several communities. Yet, over a year-long period of ritual feasting, only a hundred or so pigs may be killed by members of a single community. In contrast, the highly competitive, "less regulated" (in terms of scheduling) feasting of the Mae Enga of Highland New Guinea may involve the killing of up to 240 pigs per feast, and there may be several feasts per year, depending upon the productivity (and production) of feasting resources, as well as the number of aspiring leaders competing for prestige . Thus, ad hoc feasts are often geared more toward the enhancement of prestige than are feasts that are highly structured by a ritual cycle. Moreover, the feasibility of ad hoc feasts as aggrandizing vehicles may be restricted by the potential of the subsistence economy to generate a surplus . Societies with seasonally structured subsistence abundances may be restricted to feasting in the contexts of seasonally structured ritual feasts. This is especially the case in the context of the unpredictable yields of hunting and gathering and dry-farm agriculture. The scheduling of feasts associated with rite-ofpassage rituals, such as weddings and initiation rituals, and with life-crisis events, such as mourning rituals, may not be "regulated" to the same degree as feasts connected with a ritual cycle. Mourning rituals, for instance, can take place months or years after a death, allowing a family to collect the necessary food and goods. Mortuary feasts are indeed prime aggrandizing events in various parts of North America and Melanesia . Initiation rites also may be postponed until there are enough initiates (i.e., the wherewithal to hold appropriately elaborate ceremonies) . On the other hand, male initiation rites in many large-scale New Guinean societies were highly regulated in the sense of occurring at regular intervals. These ceremonies are conducted by part-time ritual specialists (hatathli) but are hosted by the patient and the patient's family. Thus, this feasting is not competitive, but an obligation of the patient and his entourage. In contrast to feasts hosted by self-selected aggrandizers, the fact that the host is "chosen" more or less randomly (i.e., whoever gets sick) creates a situation in which feasting operates to distribute rather than consolidate responsibility, and perhaps prestige, and serves as an element in a larger ritual structure that operates to maintain community and tribal integration。In contrast, in New Guinea, feasts associated with healing ceremonies can be an avenue for prestige-enhancement on the part of big men. The relatives of the "victim" may organize and finance the feast, and particularly competitive individuals can search out opportunities (i.e., sick relatives) to exploit by feasting.
The paper begins with a discussion of feasting in small-scale societies and suggests that recognizing the manner of articulation of three particular variables-the scale of participation and financing, the frequency and structure of occurrence, and the food resources used-is an effective starting point for gauging how politically charged feasts may have been, the actual social role feasting may have played, and the variation in feasting practices across space and through time. I then present the evidence for feasting in the Southwest from A.D. 850 to present, focusing particular attention on these three variables. I conclude the paper with a case study that highlights feasting as a political resource that is linked with, but subordinate to, other resources of power, especially the control of rituals and of the knowledge necessary to perform them. There are two sides to all communal behaviors: one that integrates and one that differentiates. Communal feasts are no exception. By their very nature, feasts bring people together to experience one of life's biological necessities in a communal, social manner. The communal exchange and sharing of food may be considered one of the most fundamental human transactions that, through social interaction and exchange, promotes social integration. This integrative aspect of feasting can have significant social and economic consequences. As Rappaport and Ford have noted, feasting can be an important mechanism for redistributing food among community members, and thus can be instrumental in promoting economic and social interdependence among community members. At the same time, to varying degrees, feasting can be an active context in which social hierarchy is established. Indeed, the hosting of feasts may be a particularly effective means through which to demonstrate one's economic and political abilities, and to engender prestige and the support of followers. Numerous aspects of feasting operate as public counting and ordering devices, which in turn reduce the vagueness of a social and political situation by promoting social comparison. For example, depending upon the quantity and quality of resources mobilized for communal feasts, and the frequency with which they are mobilized, feasting can be a quantitative measure of the abilities of the host as an efficient, skillful, vital, and generous leader. Since organizing and financing communal feasts often involves the cooperation of multiple individuals (e.g., communal hunts or large-scale domestic animal slaughters often precede communal feasts), large and elaborate feasts indicate that the host has control over the labor of others and is an efficient mobilizer of cooperative effort and, thus, a potential leader. Moreover, the degree to which the host successfully coordinates intercommunity participation at feasts is a particularly strong incentive for community allegiance and support, especially in situations involving supra-community alliances, exchange and marriage relations, conflict, and competition notes, "when food distributions are financed from outside,such as in many Highland New Guinea Big-man societies, abundance of food also is an indicator of a broad network of supportive social ties." Finally, if rare, exotic, or highly valued resources are presented and consumed in a feast, both the host's success as a procurer of these resources and, potentially, the host's control of sacred knowledge (i.e., the knowledge of how to acquire the exotic or rare resources) may be demonstrated by hosting a particularly successful feast.These display/performance aspects of communal feasting that promote social comparison operate in tandem with feasting's more fundamental capacity as a vehicle for engendering political support by creating debt-obligations through the unequal imparting of gifts and food). The creation of debt obligation is, in many anthropologists' eyes, a fundamental aspect of feasting that makes it a potent political vehicle . However, indebtedness may not be an effective political resource unless there are social or technological mechanisms in place that allow for the monopolization of the resources necessary to finance a feast so that a number of potential followers incur debt without the means for repayment by either paying off the debt or financing their own feast. It is only under these conditions that feasts may be used effectively to amplify and maintain social differences over the long term (e.g., a generation), regardless of the productivity of the environment. In addition, financing a feast may place the host in considerable debt to the host's supporters, a situation that may actually lessen the status of the host as an independent leader. Communal feasting also may be part of healing ceremonies, which are not usually structured by ritual cycles but which must occur on a fairly regular basis to prevent or cure illness. The Blessingway ceremony of the Navajo, for example, involves the feeding of the audience, which can sometimes number in the thousandsThese points are addressed below.There is tremendous variation in what has been termed feasting in middle-range societies. Most discussions of this variation have been couched in terms of ideal "types".The approach taken here, while not denying the utility of types as analytical constructs, focuses more on three particular dimensions of feasting that can combine in a number of ways to produce a variety of social and economic effects: 1) the scale of participation and financing; 2) the frequency and structure of occurrence; and 3) the food resources used. Communal feasts can range in scale from those in which a single household sponsors a feast for bilateral relatives or other households to those hosted by entire villages and offered to members of multiple communities. The scale of most communal feasts ranks somewhere between these two extremes. Among the Mae Enga and Chimbu of Highland New Guinea, feasts may be hosted by tribes (made up of clans), clans, subclans, or lineages. Many feasts in Puebloan communities are sponsored by multiple households within a community, with much of the community invited (or obligated) to participate . To a large extent, the means through which a feast is financed, and ultimately the purpose of the feast, determines its scale. Here it is useful to distinguish between household-level finance and supra-household finance . Assuming there are inherent limits on home production as a means of expanding the resource base of a feast, the establishment of social relations outside the lineage or household through sodality membership, marriage, exchange, or indebtedness all become paramount to hosting and financing largescale feasts. Postcontact Mae Enga big men of highland New Guinea, for example, may draw upon upwards of 2,000 followers for a single feast . The greater the scale of the feast, the greater the potential for prestige enhancement on the part of sponsors. Moreover, using feasting as a recurrent social practice to enhance prestige usually requires financing on a larger scale than that of the single household or lineage (Young 1985). Scale also may relate to whether feasts are primarily "intra-communal" or "inter-communal" affairs (although this distinction is often hard to make ethnographically.) Hayden (1995), for instance, delineates within-group feasts, which he terms "solidarity feasts," and between-group feasts, which are either "reciprocal" or "competitive" feasts. Reciprocal feasts function primarily to create alliances and promote cooperation among groups, while competitive feasts are hosted primarily to enhance the prestige, and political and economic support, of an aggrandizer and his/her followers or to negotiate and reaffirm extra-community political ties and agreement.Many within-group feasts have the principal effect of facilitating social integration or redistributing food among households within the village . Feasts of this type are usually financed by multiple households (often anonymously) and tend to be relatively non-competitive in nature. Compared to situations in which sponsors compete to enhance their prestige by contributing more to a feast than other sponsors, the inherent anonymity of, and lack of direct competition for, prestige in many intra-community integrative feasts tends to limit the quanitity of resources mobilized. In contrast, when feasts involve more than one community, cooperative effort on the part of village coresidents often facilitates the financing of larger and more elaborate feasts than is possible in intra-village contexts. It is in the context of inter-community (or large-scale) feasts that the prestige of aggrandizers becomes greatly enhanced relative to their followers.Much of the variation in occurrence of feasts is related to the degree to which feasting is regulated by the structure of ritual. Feasting often is associated with calendrical rituals that occur at prescribed times of the year. The kyanakwe feasting ceremony of the Zuni, for example, is held four times a year. Many feasts are held only once every several years. The Kaiko ritual of the Tsembaga of New Guinea may take place only once every 7 to 15 years . The scale of these feasts is smaller than that of unregulated or ad hoc feasts. Kaiko ritual feasts involve the participation of hundreds of people from several communities. Yet, over a year-long period of ritual feasting, only a hundred or so pigs may be killed by members of a single community. In contrast, the highly competitive, "less regulated" (in terms of scheduling) feasting of the Mae Enga of Highland New Guinea may involve the killing of up to 240 pigs per feast, and there may be several feasts per year, depending upon the productivity (and production) of feasting resources, as well as the number of aspiring leaders competing for prestige . Thus, ad hoc feasts are often geared more toward the enhancement of prestige than are feasts that are highly structured by a ritual cycle. Moreover, the feasibility of ad hoc feasts as aggrandizing vehicles may be restricted by the potential of the subsistence economy to generate a surplus . Societies with seasonally structured subsistence abundances may be restricted to feasting in the contexts of seasonally structured ritual feasts. This is especially the case in the context of the unpredictable yields of hunting and gathering and dry-farm agriculture. The scheduling of feasts associated with rite-ofpassage rituals, such as weddings and initiation rituals, and with life-crisis events, such as mourning rituals, may not be "regulated" to the same degree as feasts connected with a ritual cycle. Mourning rituals, for instance, can take place months or years after a death, allowing a family to collect the necessary food and goods. Mortuary feasts are indeed prime aggrandizing events in various parts of North America and Melanesia . Initiation rites also may be postponed until there are enough initiates (i.e., the wherewithal to hold appropriately elaborate ceremonies) . On the other hand, male initiation rites in many large-scale New Guinean societies were highly regulated in the sense of occurring at regular intervals. These ceremonies are conducted by part-time ritual specialists (hatathli) but are hosted by the patient and the patient's family. Thus, this feasting is not competitive, but an obligation of the patient and his entourage. In contrast to feasts hosted by self-selected aggrandizers, the fact that the host is "chosen" more or less randomly (i.e., whoever gets sick) creates a situation in which feasting operates to distribute rather than consolidate responsibility, and perhaps prestige, and serves as an element in a larger ritual structure that operates to maintain community and tribal integration。In contrast, in New Guinea, feasts associated with healing ceremonies can be an avenue for prestige-enhancement on the part of big men. The relatives of the "victim" may organize and finance the feast, and particularly competitive individuals can search out opportunities (i.e., sick relatives) to exploit by feasting.
考点1:"social comparison“推荐翻译为”社会攀比“ 考点2:"Big-man societies“必须翻译为”头人社会“,人类学专业术语,固定译法 考点3:"aggrandizer“推荐翻译为”夸富者“,该词在文中偏贬义,对于该词的翻译需要体现出“是通过财富、权利提升地位,提高自身声望和影响力”的意思;且不可译为“扩张者”(表意不明确) 考点4:" initiation rituals“推荐翻译为”成人礼“ 考点5:"victim"推荐翻译为“患者”
33e17
学术论文
社会科学
151
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: We spend a lot of time searching for things. If we know what we are looking for in advance, a memory representation of the target will be created to guide search. But if the identity of the search target is revealed simultaneously with the presentation of the search array, is a similar memory representation formed? In the present study, 96 observers determined whether a central target was present in a peripheral search array. The results revealed that as long as the central target remained available for inspection (even if only in iconic memory), observers reinspected it after each distractor was checked, apparently forgoing consolidation of the target into working memory. The present findings challenged the assumption that evaluating items in a search array must involve comparison with a template in working memory. Palmer (1995) proposed that visual search involves four processing stages: preselection, selection, postselection, and decision. Whereas many efforts have been made to examine the mechanisms of the first three stages (e.g., Lavie, Hirst, De Fockert, & Viding, 2004; Treisman & Gelade, 1980; Wolfe, 1994), little is known about the decision stage. The decision stage involves comparisons between the attentional template (a mental representation of the target) and items in the search array (to determine whether the item is a target). Standard theories of attention assume that the template comparison occurs in working memory (Bundesen, 1990; Duncan & Humphreys, 1989; Treisman & Gelade, 1980). Consistent with this assumption, evidence demonstrates that the attentional template may be stored in working memory during the initial configuration of attentional control settings (Carlisle, Arita, Pardo, & Woodman, 2011; Woodman & Arita, 2011). However, consolidating information into working memory is time consuming (Jolicœur & Dell’Acqua, 1998; Vogel, Woodman, & Luck, 2006) and may sometimes be inefficient. It has been suggested that the locations of visited distractors may not be remembered during visual search (Horowitz & Wolfe, 1998) and that the attributes of a just-attended object may not always be retained in working memory (Chen & Wyble, 2015). Moreover, the speed of search is faster than the speed of memory consolidation; for example, searching for a color among a group of distinctive colors can be as fast as 10 ms per item (Olds, Cowan, & Jolicoeur, 1999), but consolidating color information into working memory takes about 50 ms per item (Vogel et al., 2006). These findings suggest that properties of the search array need not be consolidated into working memory. Then what about the target itself? In the present study, we investigated whether target identity is stored in working memory when the identity of the target is both revealed simultaneously with the search array and remains available for inspection during search. Objects are identified and recognized more rapidly than information is consolidated into working memory (Jolicœur & Dell’Acqua, 1998; Potter, 1976; Thorpe, Fize, & Marlot, 1996). If the target identity is available for inspection only while the search array is present, observers may proceed immediately to search without stopping to consolidate the target in working memory. One way to avoid the need for consolidation is to revisit the target during search. To test the above hypothesis, we first compared search efficiency in a target-in-memory condition (in which the target identity has to be stored in working memory) with search efficiency in a target-on-screen condition (in which the target identity is shown simultaneously with the search array and remains available during the search). Wolfe, Horowitz, Kenner, Hyle, and Vasan (2004) showed that when the target identity is revealed (as briefly as 50 ms) before the presentation of the search array, an attentional template is formed to guide search and reduce the search time. To control this top-down guidance in the target-in-memory condition, we included a target-first condition in which the target identity is presented 50 ms before the search array and remains on screen until response. The results from Experiment 1 are consistent with the target-reidentification hypothesis. The search slope in the target-in-memory condition was 30.2 ms per item, indicating that the time for identifying a letter was 30.2 ms. The slope doubled in the target-on-screen condition (62.5 ms per item) and the target-first condition (65.5 ms per item), which suggests that, in these two conditions, each letter required a processing time sufficient for identifying two letters. A plausible explanation is that, after checking each distractor, observers reinspected the target. To provide convergent evidence for this possibility, we manipulated the form of the central target in Experiment 2: The central target was either upright or mirror reversed and tilted. Letters in the search array were all upright. We anticipated that it might take observers longer to identify the mirror-reversed and tilted letters than the upright letters. This extra target-encoding time should affect the intercept of the search-time function only if the central target is identified once, but it may increase the search slope if the central target is reinspected after each distractor is examined. In Experiment 3, there were two types of displays: simple and complex (Fig. 1). In the simple display, the target letter was mirror reversed and tilted, while letters in the search array were upright; in the complex display, the central letter was upright, but letters in the search array were mirror reversed and tilted. In the target-on-screen condition, which display would observers search more quickly? Existing theories predict a faster search in the simple display, but the reidentification hypothesis makes an astonishing prediction, that is, search speeds for the simple and complex displays should be similar. Experiment 3 tested this prediction. Previous work using eye tracking has suggested that in naturalistic tasks, observers may preferentially refixate parts of the work space rather than committing them to working memory (Ballard, Hayhoe, & Pelz, 1995; Hayhoe, Shrivastava, Mruczek, & Pelz, 2003). Although the search slopes reported here are unlikely to involve eye movements, in Experiment 4, we used a variant of the paradigm in which the target was mostly available in iconic memory and search could be completed with exposure times of less than 200 ms (thus minimizing the risk of eye movements). For this purpose, we switched to numeric digits and used a target preexposure of 30 ms, which was immediately followed either by the search array (immediate condition) or by an annular mask and then (1,000 ms later) by the search array (delayed condition). We reasoned that in the delayed condition, observers would need to consolidate the search target into working memory, but that in the immediate condition, iconic memory might be enough to support reidentification while not causing gaze movements. The findings of Experiments 1 and 4 showed that when target identity is revealed simultaneously with the search array and remains available for inspection (even in iconic memory), the search slope is twice that as when the target identity is stored in working memory before the search begins. In the no-working-memory (unconsolidated-target) search conditions, observers seemed to jump right into the search, depending on reidentification of the target rather than on first establishing a representation of the search target in working memory. This could explain the drop in search-rate efficiency. In a supplementary study (Experiment S2 in the Supplemental Material), the target-in-memory condition of Experiment 1 was modified so that the previewed target would always reappear at the center of the search array. Search slopes in this hybrid condition remained low, which implies that the observers, rather than being lazy, still preferred a faster search rate, as long as sufficient time was allowed to prepare the target representation (in working memory) to guide attention. The findings of Experiments 2 and 3 were consistent with the reidentification hypothesis—that is, when target identity is revealed simultaneously with the presentation of the search array and remains available for inspection, observers will reidentify the target after examining each distractor rather than taking the time to first consolidate the target identity into working memory. The present findings cannot be attributed to factors related to eye movements because the reidentification process can occur even in iconic memory and within a duration insufficient for initiating saccadic eye movements. If the reidentification hypothesis is correct, that implies that comparison between the target template and items in the search array does not necessarily occur in working memory and that visual search has an unconsolidatedtarget (working-memory-free) search mode. Horowitz and Wolfe (1998) showed that search efficiency does not change when the locations of the items in the search array randomly shift every 111 ms compared with when the items’ locations are fixed. They interpreted their findings as showing that the locations of visited items are not remembered (i.e., sampling with replacement; see also Gibson, Li, Skow, Brown, & Cooke, 2000; Gilchrist & Harvey, 2000; Horowitz & Wolfe, 2001, 2003). In fact, the doubled search slope in the nonconsolidated conditions of the present study is reminiscent of findings of this memory-free search. However, such searches have been reported in cases in which target consolidation has clearly occurred and the searched locations are not being consolidated. Some authors have indeed demonstrated conditions in which search occurs without replacement (Kristjánsson, 2000; Lleras, Rensink, & Enns, 2005; Peterson, Kramer, Wang, Irwin, & McCarley, 2001; Takeda, 2004). In the present case, it could be argued that ongoing attempts at consolidating the target interfered with searchlocation consolidation. However, the results of Experiment 3 seem to contradict this account. If searches in the target-on-screen condition did not include target reidentification, it is difficult to understand how search speed in the simple display was not faster than that in the complex display. Although target reidentification results in a steeper search slope, this strategy eliminates the time normally used for consolidating the target identity, which might involve several hundred milliseconds. Therefore, the overall efficiency of the unconsolidated-target search is likely better than that of the consolidated-target search when there is minimal preview time. Moreover, the benefit of this reidentification strategy is actually consistent with the finding that observers miss changes between two simultaneous arrays, just as they do in a flicker paradigm, but that their performance improves significantly in the simultaneous condition (e.g., ScottBrown, Baker, & Orbach, 2000). In natural tasks (e.g., making a sandwich), observers may frequently go back to check the model (Ballard et al., 1995; Hayhoe et al., 2003), similar to the target reidentification suggested by the present findings. Although eye movements are clearly involved in those natural tasks, the phenomenon observed in the present study is likely attributable to covert attention, because the findings of Experiment 4 show that the reidentification of the target could occur in iconic memory. It also implies that the comparison between the target template and items in the search array may occur either in iconic memory (Sperling, 1960) or in fragile short-term memory (Sligte, Scholte, & Lamme, 2008, 2009).
We spend a lot of time searching for things. If we know what we are looking for in advance, a memory representation of the target will be created to guide search. But if the identity of the search target is revealed simultaneously with the presentation of the search array, is a similar memory representation formed? In the present study, 96 observers determined whether a central target was present in a peripheral search array. The results revealed that as long as the central target remained available for inspection (even if only in iconic memory), observers reinspected it after each distractor was checked, apparently forgoing consolidation of the target into working memory. The present findings challenged the assumption that evaluating items in a search array must involve comparison with a template in working memory. Palmer (1995) proposed that visual search involves four processing stages: preselection, selection, postselection, and decision. Whereas many efforts have been made to examine the mechanisms of the first three stages (e.g., Lavie, Hirst, De Fockert, & Viding, 2004; Treisman & Gelade, 1980; Wolfe, 1994), little is known about the decision stage. The decision stage involves comparisons between the attentional template (a mental representation of the target) and items in the search array (to determine whether the item is a target). Standard theories of attention assume that the template comparison occurs in working memory (Bundesen, 1990; Duncan & Humphreys, 1989; Treisman & Gelade, 1980). Consistent with this assumption, evidence demonstrates that the attentional template may be stored in working memory during the initial configuration of attentional control settings (Carlisle, Arita, Pardo, & Woodman, 2011; Woodman & Arita, 2011). However, consolidating information into working memory is time consuming (Jolicœur & Dell’Acqua, 1998; Vogel, Woodman, & Luck, 2006) and may sometimes be inefficient. It has been suggested that the locations of visited distractors may not be remembered during visual search (Horowitz & Wolfe, 1998) and that the attributes of a just-attended object may not always be retained in working memory (Chen & Wyble, 2015). Moreover, the speed of search is faster than the speed of memory consolidation; for example, searching for a color among a group of distinctive colors can be as fast as 10 ms per item (Olds, Cowan, & Jolicoeur, 1999), but consolidating color information into working memory takes about 50 ms per item (Vogel et al., 2006). These findings suggest that properties of the search array need not be consolidated into working memory. Then what about the target itself? In the present study, we investigated whether target identity is stored in working memory when the identity of the target is both revealed simultaneously with the search array and remains available for inspection during search. Objects are identified and recognized more rapidly than information is consolidated into working memory (Jolicœur & Dell’Acqua, 1998; Potter, 1976; Thorpe, Fize, & Marlot, 1996). If the target identity is available for inspection only while the search array is present, observers may proceed immediately to search without stopping to consolidate the target in working memory. One way to avoid the need for consolidation is to revisit the target during search. To test the above hypothesis, we first compared search efficiency in a target-in-memory condition (in which the target identity has to be stored in working memory) with search efficiency in a target-on-screen condition (in which the target identity is shown simultaneously with the search array and remains available during the search). Wolfe, Horowitz, Kenner, Hyle, and Vasan (2004) showed that when the target identity is revealed (as briefly as 50 ms) before the presentation of the search array, an attentional template is formed to guide search and reduce the search time. To control this top-down guidance in the target-in-memory condition, we included a target-first condition in which the target identity is presented 50 ms before the search array and remains on screen until response. The results from Experiment 1 are consistent with the target-reidentification hypothesis. The search slope in the target-in-memory condition was 30.2 ms per item, indicating that the time for identifying a letter was 30.2 ms. The slope doubled in the target-on-screen condition (62.5 ms per item) and the target-first condition (65.5 ms per item), which suggests that, in these two conditions, each letter required a processing time sufficient for identifying two letters. A plausible explanation is that, after checking each distractor, observers reinspected the target. To provide convergent evidence for this possibility, we manipulated the form of the central target in Experiment 2: The central target was either upright or mirror reversed and tilted. Letters in the search array were all upright. We anticipated that it might take observers longer to identify the mirror-reversed and tilted letters than the upright letters. This extra target-encoding time should affect the intercept of the search-time function only if the central target is identified once, but it may increase the search slope if the central target is reinspected after each distractor is examined. In Experiment 3, there were two types of displays: simple and complex (Fig. 1). In the simple display, the target letter was mirror reversed and tilted, while letters in the search array were upright; in the complex display, the central letter was upright, but letters in the search array were mirror reversed and tilted. In the target-on-screen condition, which display would observers search more quickly? Existing theories predict a faster search in the simple display, but the reidentification hypothesis makes an astonishing prediction, that is, search speeds for the simple and complex displays should be similar. Experiment 3 tested this prediction. Previous work using eye tracking has suggested that in naturalistic tasks, observers may preferentially refixate parts of the work space rather than committing them to working memory (Ballard, Hayhoe, & Pelz, 1995; Hayhoe, Shrivastava, Mruczek, & Pelz, 2003). Although the search slopes reported here are unlikely to involve eye movements, in Experiment 4, we used a variant of the paradigm in which the target was mostly available in iconic memory and search could be completed with exposure times of less than 200 ms (thus minimizing the risk of eye movements). For this purpose, we switched to numeric digits and used a target preexposure of 30 ms, which was immediately followed either by the search array (immediate condition) or by an annular mask and then (1,000 ms later) by the search array (delayed condition). We reasoned that in the delayed condition, observers would need to consolidate the search target into working memory, but that in the immediate condition, iconic memory might be enough to support reidentification while not causing gaze movements. The findings of Experiments 1 and 4 showed that when target identity is revealed simultaneously with the search array and remains available for inspection (even in iconic memory), the search slope is twice that as when the target identity is stored in working memory before the search begins. In the no-working-memory (unconsolidated-target) search conditions, observers seemed to jump right into the search, depending on reidentification of the target rather than on first establishing a representation of the search target in working memory. This could explain the drop in search-rate efficiency. In a supplementary study (Experiment S2 in the Supplemental Material), the target-in-memory condition of Experiment 1 was modified so that the previewed target would always reappear at the center of the search array. Search slopes in this hybrid condition remained low, which implies that the observers, rather than being lazy, still preferred a faster search rate, as long as sufficient time was allowed to prepare the target representation (in working memory) to guide attention. The findings of Experiments 2 and 3 were consistent with the reidentification hypothesis—that is, when target identity is revealed simultaneously with the presentation of the search array and remains available for inspection, observers will reidentify the target after examining each distractor rather than taking the time to first consolidate the target identity into working memory. The present findings cannot be attributed to factors related to eye movements because the reidentification process can occur even in iconic memory and within a duration insufficient for initiating saccadic eye movements. If the reidentification hypothesis is correct, that implies that comparison between the target template and items in the search array does not necessarily occur in working memory and that visual search has an unconsolidatedtarget (working-memory-free) search mode. Horowitz and Wolfe (1998) showed that search efficiency does not change when the locations of the items in the search array randomly shift every 111 ms compared with when the items’ locations are fixed. They interpreted their findings as showing that the locations of visited items are not remembered (i.e., sampling with replacement; see also Gibson, Li, Skow, Brown, & Cooke, 2000; Gilchrist & Harvey, 2000; Horowitz & Wolfe, 2001, 2003). In fact, the doubled search slope in the nonconsolidated conditions of the present study is reminiscent of findings of this memory-free search. However, such searches have been reported in cases in which target consolidation has clearly occurred and the searched locations are not being consolidated. Some authors have indeed demonstrated conditions in which search occurs without replacement (Kristjánsson, 2000; Lleras, Rensink, & Enns, 2005; Peterson, Kramer, Wang, Irwin, & McCarley, 2001; Takeda, 2004). In the present case, it could be argued that ongoing attempts at consolidating the target interfered with searchlocation consolidation. However, the results of Experiment 3 seem to contradict this account. If searches in the target-on-screen condition did not include target reidentification, it is difficult to understand how search speed in the simple display was not faster than that in the complex display. Although target reidentification results in a steeper search slope, this strategy eliminates the time normally used for consolidating the target identity, which might involve several hundred milliseconds. Therefore, the overall efficiency of the unconsolidated-target search is likely better than that of the consolidated-target search when there is minimal preview time. Moreover, the benefit of this reidentification strategy is actually consistent with the finding that observers miss changes between two simultaneous arrays, just as they do in a flicker paradigm, but that their performance improves significantly in the simultaneous condition (e.g., ScottBrown, Baker, & Orbach, 2000). In natural tasks (e.g., making a sandwich), observers may frequently go back to check the model (Ballard et al., 1995; Hayhoe et al., 2003), similar to the target reidentification suggested by the present findings. Although eye movements are clearly involved in those natural tasks, the phenomenon observed in the present study is likely attributable to covert attention, because the findings of Experiment 4 show that the reidentification of the target could occur in iconic memory. It also implies that the comparison between the target template and items in the search array may occur either in iconic memory (Sperling, 1960) or in fragile short-term memory (Sligte, Scholte, & Lamme, 2008, 2009).
考点1:“the identity of the search target”推荐翻译为“搜索目标本身” 考点2:“iconic memory”应翻译为“映像记忆” 考点3:“preselection, selection, postselection, and decision”推荐翻译为“选择前阶段、选择阶段、选择后阶段和决策阶段” 考点4:“visited distractors”应翻译为“被注意过的干扰项/被检视过的干扰项”,不能翻译为“已访问的干扰项” 考点5:“target-in-memory condition”推荐翻译为“目标在记忆中的条件” 考点6:“target-on-screen condition”推荐翻译为“目标在屏幕上的条件” 考点7:“eye movements”推荐翻译为“眼动” 考点8:“reidentification hypothesis”推荐译为“再识别假设”或“重新识别假设”
346e2
学术论文
社会科学
134
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 电化学系统:能量转换与反应动力学的解析 电化学是研究电能与化学能相互转换规律的学科,其本质是氧化还原反应中电子的定向转移过程。这类反应通过两类基础装置实现:原电池将化学能自发转化为电能,而电解池则消耗电能驱动非自发化学反应。 一、电池系统的结构与功能分类 在原电池中,负极(亦称阳极)发生氧化反应释放电子,正极(亦称阴极)发生还原反应接收电子,电子流经外电路形成电流。典型如锌铜电池: 1.锌阳极发生氧化反应(Zn → Zn²⁺ + 2e⁻) 2.铜阴极发生还原反应(Cu²⁺ + 2e⁻ → Cu) 3.两类特殊电池拓展了应用场景: 4.蓄电池(二次电池):通过可逆反应实现反复充放电,其循环寿命依赖于电极表面形成的固体电解质界面膜稳定性,该膜可抑制副反应并保护电极结构; 5.贮备电池:采用物理隔离技术使电极与电解质长期分离(如电解质固态化),激活后瞬时放电,适用于应急设备与军事领域。 电池内部结构中,隔膜承担双重使命:阻隔正负极接触防止短路,同时允许离子穿透以维持电荷平衡。电解液作为离子传导介质,其组成直接影响离子迁移率及电池内阻。近年兴起的固态电解质技术摒弃传统液态体系,利用无机晶体或聚合物传导离子,显著提升安全性并降低欧姆内阻。 二、电极系统与界面控制机制 电化学研究依赖三电极体系精确测量: 1.工作电极:承载目标反应的核心平台,例如研究二氧化碳还原的铂片电极; 2.对电极(辅助电极):与工作电极构成电流回路,通常选用化学惰性的铂或石墨材料; 3.参比电极:具有稳定且已知电位的基准电极(如饱和甘汞电极),用于测定工作电极的绝对电位值。 在电极/溶液界面处,带相反电荷的粒子定向排布形成双电层,该结构如同微观电容器,直接影响电子转移步骤的能垒。通过对电极表面进行分子设计可制备修饰电极,例如将金属纳米粒子负载于碳基体增强电催化活性,或固定生物酶实现特异性检测。 三、反应动力学的多步骤控制 电化学反应速度由串联步骤共同决定: 1.液相传质步骤:反应物从溶液本体扩散至电极表面附近; 2.前置转化:反应粒子在电极界面发生预活化(如有机分子的取向吸附或质子化); 3.电子转移步骤:电子穿越双电层的核心电荷传递过程; 4.随后转化:初生产物的脱附、重排或二次化学反应。 当外电路电流通过时,电极电位将偏离平衡状态,此现象称为极化。根据起源可分为: 1.电化学极化:电荷传递步骤动力学迟滞导致; 2.浓差极化:反应物消耗或产物积累引发的浓度梯度效应。 两者共同构成极化内阻,其数值可通过扫描电位绘制的极化曲线量化分析。曲线特征显示: 3.阴极极化驱使电位向负方向偏移(典型如氧气还原反应受阻); 4.阳极极化推动电位向正方向移动(如金属阳极溶解迟滞)。 过电位作为极化的直接表征,其数值与电流密度满足半对数关系(塔菲尔方程)。降低过电位需协同优化离子迁移率与电化学活性表面积,后者反映电极实际参与反应的有效界面比例。 四、能量效率的制约因素与优化策略 电流效率衡量有效电子利用率,例如电解水制氢体系中,阴极析氢反应的电流效率可达95%以上。提升效率需抑制竞争反应: 1.引入离子交换膜实现离子选择性透过(如氯碱工业阻隔氢氧根反渗); 2.调控电解液组成削弱化学极化(由前置或随后转化步骤控制的动力学阻力)。 欧姆极化源于电流流经导体时的固有电阻,其物理本质为欧姆内阻,包含电极体电阻、集流体接触电阻及电解质阻抗三部分。通过交流阻抗谱可分离识别欧姆损耗与极化损耗。 五、标准体系与理论基准框架 标准电极电位在25°C、1 mol/L浓度的理想状态下测定,以标准氢电极(铂黑电极在1 atm氢气与1 mol/L H⁺溶液中的体系)为零电位基准。例如铜电极的标准电位为+0.34 V,预示其还原倾向强于氢电极。该参数虽可预测反应方向,但实际电位受界面吸附状态及双电层结构显著影响。 电解质导电机理基于电离与电迁移:强电解质(如氯化钠)在水溶液中完全解离成正负离子,弱电解质(如醋酸)部分解离;在外加电场驱动下,离子定向电迁移形成电流,迁移速率与电场强度及离子迁移率成正比。 六、技术前沿与工程应用突破 现代电化学技术正在重塑能源格局: 1.全固态电池采用陶瓷固态电解质替代有机电解液,消除燃爆风险并提升能量密度; 2.液流电池储能系统通过离子交换膜分隔正负极电解液,实现百兆瓦级可再生能源存储; 3.电化学合成利用修饰电极的高选择性,将二氧化碳转化为甲酸或乙烯等高值化学品。 4.降低极化损失仍是核心挑战。以质子交换膜燃料电池为例: 5.增加铂催化剂的电化学活性表面积可加速迟缓的氧还原反应; 6.优化乙醇氧化的前置转化路径(如C-C键断裂步骤)能提升燃料电池效率; 7.调控固体电解质界面膜组分可减少锂离子电池首次充放电的容量损失。 8.在环境治理领域,电解技术展现出独特优势: 9.电化学高级氧化通过阳极反应产生活性氧自由基降解有机污染物; 10.电沉积法回收废水中的铜、镍等重金属,电流效率达80%以上; 11.电解池耦合离子交换膜实现盐差能发电,开辟蓝色能源新途径。 七、表征技术与过程监控 电化学动力学研究依赖先进原位表征: 1.极化曲线定量揭示电极过程的速率控制步骤; 2.交流阻抗谱解析电池内阻组成(区分欧姆内阻与极化内阻); 3.显微技术观测固体电解质界面膜的形貌演化; 4.光谱分析追踪吸附中间体的生成与转化。 在线监控参数包括: 1.电流效率实时反映反应选择性; 2.过电位监控系统能耗状态; 3.电化学活性表面积评估催化剂老化程度。 结论 电化学系统的核心在于协调电子转移步骤的热力学驱动力与动力学限制。从原电池自主供能到电解池强制反应,从蓄电池的循环再生再到贮备电池的长效待机,技术创新始终围绕克服欧姆极化与浓差极化、提升电流效率展开。标准电极电位奠定了理论预测基础,参比电极提供了实验测量基准,而修饰电极与固态电解质等新材料则持续突破性能边界。未来,通过多尺度调控液相传质步骤、前置转化路径及界面双电层结构,电化学技术将在清洁能源存储、污染物深度治理及高端化学品合成领域开启新纪元。
电化学系统:能量转换与反应动力学的解析 电化学是研究电能与化学能相互转换规律的学科,其本质是氧化还原反应中电子的定向转移过程。这类反应通过两类基础装置实现:原电池将化学能自发转化为电能,而电解池则消耗电能驱动非自发化学反应。 一、电池系统的结构与功能分类 在原电池中,负极(亦称阳极)发生氧化反应释放电子,正极(亦称阴极)发生还原反应接收电子,电子流经外电路形成电流。典型如锌铜电池: 1.锌阳极发生氧化反应(Zn → Zn²⁺ + 2e⁻) 2.铜阴极发生还原反应(Cu²⁺ + 2e⁻ → Cu) 3.两类特殊电池拓展了应用场景: 4.蓄电池(二次电池):通过可逆反应实现反复充放电,其循环寿命依赖于电极表面形成的固体电解质界面膜稳定性,该膜可抑制副反应并保护电极结构; 5.贮备电池:采用物理隔离技术使电极与电解质长期分离(如电解质固态化),激活后瞬时放电,适用于应急设备与军事领域。 电池内部结构中,隔膜承担双重使命:阻隔正负极接触防止短路,同时允许离子穿透以维持电荷平衡。电解液作为离子传导介质,其组成直接影响离子迁移率及电池内阻。近年兴起的固态电解质技术摒弃传统液态体系,利用无机晶体或聚合物传导离子,显著提升安全性并降低欧姆内阻。 二、电极系统与界面控制机制 电化学研究依赖三电极体系精确测量: 1.工作电极:承载目标反应的核心平台,例如研究二氧化碳还原的铂片电极; 2.对电极(辅助电极):与工作电极构成电流回路,通常选用化学惰性的铂或石墨材料; 3.参比电极:具有稳定且已知电位的基准电极(如饱和甘汞电极),用于测定工作电极的绝对电位值。 在电极/溶液界面处,带相反电荷的粒子定向排布形成双电层,该结构如同微观电容器,直接影响电子转移步骤的能垒。通过对电极表面进行分子设计可制备修饰电极,例如将金属纳米粒子负载于碳基体增强电催化活性,或固定生物酶实现特异性检测。 三、反应动力学的多步骤控制 电化学反应速度由串联步骤共同决定: 1.液相传质步骤:反应物从溶液本体扩散至电极表面附近; 2.前置转化:反应粒子在电极界面发生预活化(如有机分子的取向吸附或质子化); 3.电子转移步骤:电子穿越双电层的核心电荷传递过程; 4.随后转化:初生产物的脱附、重排或二次化学反应。 当外电路电流通过时,电极电位将偏离平衡状态,此现象称为极化。根据起源可分为: 1.电化学极化:电荷传递步骤动力学迟滞导致; 2.浓差极化:反应物消耗或产物积累引发的浓度梯度效应。 两者共同构成极化内阻,其数值可通过扫描电位绘制的极化曲线量化分析。曲线特征显示: 3.阴极极化驱使电位向负方向偏移(典型如氧气还原反应受阻); 4.阳极极化推动电位向正方向移动(如金属阳极溶解迟滞)。 过电位作为极化的直接表征,其数值与电流密度满足半对数关系(塔菲尔方程)。降低过电位需协同优化离子迁移率与电化学活性表面积,后者反映电极实际参与反应的有效界面比例。 四、能量效率的制约因素与优化策略 电流效率衡量有效电子利用率,例如电解水制氢体系中,阴极析氢反应的电流效率可达95%以上。提升效率需抑制竞争反应: 1.引入离子交换膜实现离子选择性透过(如氯碱工业阻隔氢氧根反渗); 2.调控电解液组成削弱化学极化(由前置或随后转化步骤控制的动力学阻力)。 欧姆极化源于电流流经导体时的固有电阻,其物理本质为欧姆内阻,包含电极体电阻、集流体接触电阻及电解质阻抗三部分。通过交流阻抗谱可分离识别欧姆损耗与极化损耗。 五、标准体系与理论基准框架 标准电极电位在25°C、1 mol/L浓度的理想状态下测定,以标准氢电极(铂黑电极在1 atm氢气与1 mol/L H⁺溶液中的体系)为零电位基准。例如铜电极的标准电位为+0.34 V,预示其还原倾向强于氢电极。该参数虽可预测反应方向,但实际电位受界面吸附状态及双电层结构显著影响。 电解质导电机理基于电离与电迁移:强电解质(如氯化钠)在水溶液中完全解离成正负离子,弱电解质(如醋酸)部分解离;在外加电场驱动下,离子定向电迁移形成电流,迁移速率与电场强度及离子迁移率成正比。 六、技术前沿与工程应用突破 现代电化学技术正在重塑能源格局: 1.全固态电池采用陶瓷固态电解质替代有机电解液,消除燃爆风险并提升能量密度; 2.液流电池储能系统通过离子交换膜分隔正负极电解液,实现百兆瓦级可再生能源存储; 3.电化学合成利用修饰电极的高选择性,将二氧化碳转化为甲酸或乙烯等高值化学品。 4.降低极化损失仍是核心挑战。以质子交换膜燃料电池为例: 5.增加铂催化剂的电化学活性表面积可加速迟缓的氧还原反应; 6.优化乙醇氧化的前置转化路径(如C-C键断裂步骤)能提升燃料电池效率; 7.调控固体电解质界面膜组分可减少锂离子电池首次充放电的容量损失。 8.在环境治理领域,电解技术展现出独特优势: 9.电化学高级氧化通过阳极反应产生活性氧自由基降解有机污染物; 10.电沉积法回收废水中的铜、镍等重金属,电流效率达80%以上; 11.电解池耦合离子交换膜实现盐差能发电,开辟蓝色能源新途径。 七、表征技术与过程监控 电化学动力学研究依赖先进原位表征: 1.极化曲线定量揭示电极过程的速率控制步骤; 2.交流阻抗谱解析电池内阻组成(区分欧姆内阻与极化内阻); 3.显微技术观测固体电解质界面膜的形貌演化; 4.光谱分析追踪吸附中间体的生成与转化。 在线监控参数包括: 1.电流效率实时反映反应选择性; 2.过电位监控系统能耗状态; 3.电化学活性表面积评估催化剂老化程度。 结论 电化学系统的核心在于协调电子转移步骤的热力学驱动力与动力学限制。从原电池自主供能到电解池强制反应,从蓄电池的循环再生再到贮备电池的长效待机,技术创新始终围绕克服欧姆极化与浓差极化、提升电流效率展开。标准电极电位奠定了理论预测基础,参比电极提供了实验测量基准,而修饰电极与固态电解质等新材料则持续突破性能边界。未来,通过多尺度调控液相传质步骤、前置转化路径及界面双电层结构,电化学技术将在清洁能源存储、污染物深度治理及高端化学品合成领域开启新纪元。
考点1:“蓄电池(二次电池)“必须译为:”:“Accumulator“ 考点2:“固体电解质界面膜“必须译为:”Solid Electrolyte Interphase“ 考点3:“工作电极“推荐译为:”Working Electrode“ 考点4:“对电极(辅助电极)“推荐译为:”Counter Electrode“ 考点5:“双电层“推荐译为:”Double Layer“ 考点6:“液相传质步骤“推荐译为:”Liquid Transfer Procedure“ 考点7:“前置转化“推荐译为:”Preceding Reaction“ 考点8:“浓差极化“推荐译为:”Concentration Polarization“ 考点9:离子迁移率 推荐译为”Ion Migration Rate“ 考点10:“固态电解质”必须译为 “solid-state electrolyte”
364ea
学术论文
自然科学
121
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 一、基于用户信息的基本协同过滤    基于用户信息的基本协同过滤是抖音整个算法体系中最基础和最简单的算法,也是在视频推广过程中普遍应用的算法。抖音通过获取用户注册时的基本信息,如性别、年龄、地址和基本兴趣点,对用户的画像有了大致的描绘。其后,在信息分发过程中,抖音通过考虑用户之间的相似程度进行相似内容的推荐。当用户开始接触平台,且所提供的信息越详细或越准确,其对用户需求的判断越接近用户的真实需要。我们可以构建如下推荐模式模型。    假如A、B、C、D的基本信息较为相似,则前期在A、B、C共同感兴趣的内容或产生点赞、评论、转发等使用行为的内容会较为优先推荐给D。比如,抖音的主界面分为同城和推荐两个模块,推荐模块一个重要的原则就是视频内容和用户兴趣的匹配程度。匹配程度越高的信息越能够被推荐。通常,在平台用户的初期使用阶段,此种获取方式最为主要,这种相对简单的算法推荐原则对于用户具体兴趣的判断是相对模糊的,更多的是相似特征人群的兴趣集合判断。可以说,不断扩大的用户数量和使用行为为这类基础算法提供了数据源,通过海量数据的收集与分析,这类算法的精准程度会不断提升。与快手等其他视频应用不同,抖音用户较为集中在城市,用户的学历和年龄差距相对较小,媒介素养也较为相似,而这也是这类基础算法较为适用的重要原因。    二、基于“去中心化”的精准推送    社交媒体最重要的原则就是“去中心化”,“把关人”的作用逐渐弱化,每个用户都是传播场域中的节点,每个节点都可以独立地生产内容,拥有一定的话语权,内容并非集中于少数的关键用户。在这种“去中心化”的精准推送中,内容和社交关系成为被抓取作为信息精准分发的主要依据。我们可以通过用户个人的视角构建如下信息获得模型。    这一类的精准算法可以分为两大类:一类是以内容兴趣点为筛选维度的推荐,这个维度下,现实社交环境中的联系较弱,甚至没有联系;另一类是以社交强联系为筛选维度的推荐,在通讯录中的好友、同学都会成为推荐所抓取的对象。重合部分往往会获得较多的推荐。    内容维度可以理解为以相同的职业、爱好、话题组成的相近兴趣的集合。平台基于用户使用行为的习惯,诸如点赞、评论、转发等行为,通过数据驱动的方式挖掘用户的兴趣点,从而在下一轮的视频推荐中合理选择,进一步取悦用户,增加黏度。    社交关系维度可以理解为以现实的社交关系为连接的集合。这类关系基于社交关系的联系程度从而变得更强。每个人的社交圈都随着移动互联技术的进步而不断扩大,但是,用户本身和现实社交中好友或同学的关注点相似性不高。正是基于这种算法推荐,用户在抖音的使用过程中通常会看到通讯录好友发布的内容,因为平台通过算法也优先推荐强关系的好友发布的内容。这反映出短视频平台重要的社交属性,通过自己拍摄视频在强关系中的展示,能够满足使用者的需要,增强认同感和满足被别人了解的社交需求。
一、基于用户信息的基本协同过滤    基于用户信息的基本协同过滤是抖音整个算法体系中最基础和最简单的算法,也是在视频推广过程中普遍应用的算法。抖音通过获取用户注册时的基本信息,如性别、年龄、地址和基本兴趣点,对用户的画像有了大致的描绘。其后,在信息分发过程中,抖音通过考虑用户之间的相似程度进行相似内容的推荐。当用户开始接触平台,且所提供的信息越详细或越准确,其对用户需求的判断越接近用户的真实需要。我们可以构建如下推荐模式模型。    假如A、B、C、D的基本信息较为相似,则前期在A、B、C共同感兴趣的内容或产生点赞、评论、转发等使用行为的内容会较为优先推荐给D。比如,抖音的主界面分为同城和推荐两个模块,推荐模块一个重要的原则就是视频内容和用户兴趣的匹配程度。匹配程度越高的信息越能够被推荐。通常,在平台用户的初期使用阶段,此种获取方式最为主要,这种相对简单的算法推荐原则对于用户具体兴趣的判断是相对模糊的,更多的是相似特征人群的兴趣集合判断。可以说,不断扩大的用户数量和使用行为为这类基础算法提供了数据源,通过海量数据的收集与分析,这类算法的精准程度会不断提升。与快手等其他视频应用不同,抖音用户较为集中在城市,用户的学历和年龄差距相对较小,媒介素养也较为相似,而这也是这类基础算法较为适用的重要原因。    二、基于“去中心化”的精准推送    社交媒体最重要的原则就是“去中心化”,“把关人”的作用逐渐弱化,每个用户都是传播场域中的节点,每个节点都可以独立地生产内容,拥有一定的话语权,内容并非集中于少数的关键用户。在这种“去中心化”的精准推送中,内容和社交关系成为被抓取作为信息精准分发的主要依据。我们可以通过用户个人的视角构建如下信息获得模型。    这一类的精准算法可以分为两大类:一类是以内容兴趣点为筛选维度的推荐,这个维度下,现实社交环境中的联系较弱,甚至没有联系;另一类是以社交强联系为筛选维度的推荐,在通讯录中的好友、同学都会成为推荐所抓取的对象。重合部分往往会获得较多的推荐。    内容维度可以理解为以相同的职业、爱好、话题组成的相近兴趣的集合。平台基于用户使用行为的习惯,诸如点赞、评论、转发等行为,通过数据驱动的方式挖掘用户的兴趣点,从而在下一轮的视频推荐中合理选择,进一步取悦用户,增加黏度。    社交关系维度可以理解为以现实的社交关系为连接的集合。这类关系基于社交关系的联系程度从而变得更强。每个人的社交圈都随着移动互联技术的进步而不断扩大,但是,用户本身和现实社交中好友或同学的关注点相似性不高。正是基于这种算法推荐,用户在抖音的使用过程中通常会看到通讯录好友发布的内容,因为平台通过算法也优先推荐强关系的好友发布的内容。这反映出短视频平台重要的社交属性,通过自己拍摄视频在强关系中的展示,能够满足使用者的需要,增强认同感和满足被别人了解的社交需求。
考点1:“基于用户信息的基本协同过滤”推荐译为“User-based Collaborative Filtering” 考点2:“视频推广”推荐译为“Video Promotion” 考点3:“基本兴趣点”推荐译为“user interest tags” 考点4:“媒介素养”推荐译为“Media Literacy” 考点5:“精准算法”推荐译为“accurate algorithm” 考点6:“增加黏度”推荐译为“increasing user stickiness” 考点7:“强关系”推荐译为“within their close circles”或“strong ties” 考点8:“认同感”推荐译为“personal validation” 考点9:“用户的画像”推荐译为“user personas” 考点10:“抓取”推荐译为“extracted” 考点11:“取悦用户”推荐译为“to better cater to the user” 考点12:“同城和推荐两个模块”中的“同城”需译为“Same City” 考点13:“同城和推荐两个模块”中的“推荐”需译为“For You”
36cee
学术论文
应用学科
13
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 自从那个下午我无意中进了这园子,就再没长久地离开过它。我一下子就理解了它的意图。正如我在一篇小说中所说的:“在人口密聚的城市里,有这样一个宁静的去处,像是上帝的苦心安排。” 两条腿残废后的最初几年,我找不到工作,找不到去路,忽然间几乎什么都找不到了,我就摇了轮椅总是到它那儿去,仅为着那儿是可以逃避一个世界的另一个世界。我在那篇小说中写道:“没处可去我便一天到晚耗在这园子里。跟上班下班一样,别人去上班我就摇了轮椅到这儿来。园子无人看管,上下班时间有些抄近路的人们从园中穿过,园子里活跃一阵,过后便沉寂下来。”“园墙在金晃晃的空气中斜切下一溜荫凉,我把轮椅开进去,把椅背放倒,坐着或是躺着,看书或者想事,撅一杈树枝左右拍打,驱赶那些和我一样不明白为什么要来这世上的小昆虫。”“蜂儿如一朵小雾稳稳地停在半空;蚂蚁摇头晃脑捋着触须,猛然间想透了什么,转身疾行而去;瓢虫爬得不耐烦了,累了祈祷一回便支开翅膀,忽悠一下升空了;树干上留着一只蝉蜕,寂寞如一间空屋;露水在草叶上滚动、聚集,压弯了草叶轰然坠地摔开万道金光。”“满园子都是草木竞相生长弄出的响动,窸窸窣窣窸窸窣窣片刻不息。”这都是真实的记录,园子荒芜但并不衰败。 除去几座殿堂我无法进去,除去那座祭坛我不能上去而只能从各个角度张望它,地坛的每一棵树下我都去过,差不多它的每一米草地上都有过我的车轮印。无论是什么季节,什么天气,什么时间,我都在这园子里呆过。有时候呆一会儿就回家,有时候就呆到满地上都亮起月光。记不清都是在它的哪些角落里了。我一连几小时专心致志地想关于死的事,也以同样的耐心和方式想过我为什么要出生。这样想了好几年,最后事情终于弄明白了:一个人,出生了,这就不再是一个可以辩论的问题,而只是上帝交给他的一个事实;上帝在交给我们这件事实的时候,已经顺便保证了它的结果,所以死是一件不必急于求成的事,死是一个必然会降临的节日。这样想过之后我安心多了,眼前的一切不再那么可怕。比如你起早熬夜准备考试的时候,忽然想起有一个长长的假期在前面等待你,你会不会觉得轻松一点?并且庆幸并且感激这样的安排? 剩下的就是怎样活的问题了,这却不是在某一个瞬间就能完全想透的、不是一次性能够解决的事,怕是活多久就要想它多久了,就像是伴你终生的魔鬼或恋人。所以,十五年了,我还是总得到那古园里去,去它的老树下或荒草边或颓墙旁,去默坐,去呆想,去推开耳边的嘈杂理一理纷乱的思绪,去窥看自己的心魂。十五年中,这古园的形体被不能理解它的人肆意雕琢,幸好有些东西是任谁也不能改变它的。譬如祭坛石门中的落日,寂静的光辉平铺的一刻,地上的每一个坎坷都被映照得灿烂;譬如在园中最为落寞的时间,一群雨燕便出来高歌,把天地都叫喊得苍凉;譬如冬天雪地上孩子的脚印,总让人猜想他们是谁,曾在哪儿做过些什么,然后又都到哪儿去了;譬如那些苍黑的古柏,你忧郁的时候它们镇静地站在那儿,你欣喜的时候它们依然镇静地站在那儿,它们没日没夜地站在那儿从你没有出生一直站到这个世界上又没了你的时候;譬如暴雨骤临园中,激起一阵阵灼烈而清纯的草木和泥土的气味,让人想起无数个夏天的事件;譬如秋风忽至,再有一场早霜,落叶或飘摇歌舞或坦然安卧,满园中播散着熨帖而微苦的味道。味道是最说不清楚的。味道不能写只能闻,要你身临其境去闻才能明了。味道甚至是难于记忆的,只有你又闻到它你才能记起它的全部情感和意蕴。所以我常常要到那园子里去。
自从那个下午我无意中进了这园子,就再没长久地离开过它。我一下子就理解了它的意图。正如我在一篇小说中所说的:“在人口密聚的城市里,有这样一个宁静的去处,像是上帝的苦心安排。” 两条腿残废后的最初几年,我找不到工作,找不到去路,忽然间几乎什么都找不到了,我就摇了轮椅总是到它那儿去,仅为着那儿是可以逃避一个世界的另一个世界。我在那篇小说中写道:“没处可去我便一天到晚耗在这园子里。跟上班下班一样,别人去上班我就摇了轮椅到这儿来。园子无人看管,上下班时间有些抄近路的人们从园中穿过,园子里活跃一阵,过后便沉寂下来。”“园墙在金晃晃的空气中斜切下一溜荫凉,我把轮椅开进去,把椅背放倒,坐着或是躺着,看书或者想事,撅一杈树枝左右拍打,驱赶那些和我一样不明白为什么要来这世上的小昆虫。”“蜂儿如一朵小雾稳稳地停在半空;蚂蚁摇头晃脑捋着触须,猛然间想透了什么,转身疾行而去;瓢虫爬得不耐烦了,累了祈祷一回便支开翅膀,忽悠一下升空了;树干上留着一只蝉蜕,寂寞如一间空屋;露水在草叶上滚动、聚集,压弯了草叶轰然坠地摔开万道金光。”“满园子都是草木竞相生长弄出的响动,窸窸窣窣窸窸窣窣片刻不息。”这都是真实的记录,园子荒芜但并不衰败。 除去几座殿堂我无法进去,除去那座祭坛我不能上去而只能从各个角度张望它,地坛的每一棵树下我都去过,差不多它的每一米草地上都有过我的车轮印。无论是什么季节,什么天气,什么时间,我都在这园子里呆过。有时候呆一会儿就回家,有时候就呆到满地上都亮起月光。记不清都是在它的哪些角落里了。我一连几小时专心致志地想关于死的事,也以同样的耐心和方式想过我为什么要出生。这样想了好几年,最后事情终于弄明白了:一个人,出生了,这就不再是一个可以辩论的问题,而只是上帝交给他的一个事实;上帝在交给我们这件事实的时候,已经顺便保证了它的结果,所以死是一件不必急于求成的事,死是一个必然会降临的节日。这样想过之后我安心多了,眼前的一切不再那么可怕。比如你起早熬夜准备考试的时候,忽然想起有一个长长的假期在前面等待你,你会不会觉得轻松一点?并且庆幸并且感激这样的安排? 剩下的就是怎样活的问题了,这却不是在某一个瞬间就能完全想透的、不是一次性能够解决的事,怕是活多久就要想它多久了,就像是伴你终生的魔鬼或恋人。所以,十五年了,我还是总得到那古园里去,去它的老树下或荒草边或颓墙旁,去默坐,去呆想,去推开耳边的嘈杂理一理纷乱的思绪,去窥看自己的心魂。十五年中,这古园的形体被不能理解它的人肆意雕琢,幸好有些东西是任谁也不能改变它的。譬如祭坛石门中的落日,寂静的光辉平铺的一刻,地上的每一个坎坷都被映照得灿烂;譬如在园中最为落寞的时间,一群雨燕便出来高歌,把天地都叫喊得苍凉;譬如冬天雪地上孩子的脚印,总让人猜想他们是谁,曾在哪儿做过些什么,然后又都到哪儿去了;譬如那些苍黑的古柏,你忧郁的时候它们镇静地站在那儿,你欣喜的时候它们依然镇静地站在那儿,它们没日没夜地站在那儿从你没有出生一直站到这个世界上又没了你的时候;譬如暴雨骤临园中,激起一阵阵灼烈而清纯的草木和泥土的气味,让人想起无数个夏天的事件;譬如秋风忽至,再有一场早霜,落叶或飘摇歌舞或坦然安卧,满园中播散着熨帖而微苦的味道。味道是最说不清楚的。味道不能写只能闻,要你身临其境去闻才能明了。味道甚至是难于记忆的,只有你又闻到它你才能记起它的全部情感和意蕴。所以我常常要到那园子里去。
考点1:“园子”应译为“Ditan Park或park”,此处的园子指北京地坛公园,不可译为garden。 考点2:“殿堂”可译为“hall”。 考点3:“那座祭坛我不能上去”表达“上去”不可用climb,此文作者行动不便、坐轮椅,因此不可翻译为climb。 考点4:“石门”推荐译为“stone arch”,用gate不合适。 考点5:“镇静地”推荐译为“sedately”,不可用calmly,calmly着重描述在压力情境下保持克制、稳定的行为方式,和此处古树的状态不符。
3c88a
文学艺术
散文
1
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 摘 要: 批判性思维是促使学生深度理解物理概念并提升其解决复杂问题能力的关键因素。然而,传统教学多聚焦于批判性思维相关的分析推理技能,较少关注批判性思维的认知基础。通过分析三个典型的物理学习案例,揭示批判性思维在物理学习中的表现特征。在此基础上,基于双加工理论和抑制控制理论,提出物理学习中批判性思维的认知基础在于学生能否识别并抑制直觉优势反应,转而调动系统2进行理性思考。接着,从认知行为与神经机制两个方面梳理抑制控制与批判性思维的实证证据。最后,从抑制控制训练融入物理课堂、提供支持性的物理课堂环境、抑制控制的自我监控与实践等方面提出对物理教学的启示。 批判性思维是21世纪人才应该具备的关键能力,尤其在科学教育领域,其对于学生形成理性判断、批判评估与创新表达的能力尤为重要。习近平总书记在中国科协第十次全国代表大会上的讲话中特别指出:“要更加重视人才自主培养,更加重视科学精神、创新能力、批判性思维的培养培育。”国内外科学课程亦高度强调批判性思维的重要性,如《普通高中物理课程标准(2017年版2020年修订)》指出:“……需要培养学生能够基于事实证据和科学推理对不同观点和结论提出质疑和批判,进行检验和修正,进而提出创新性见解的能力与品质”。 然而,我国青少年的批判性思维能力发展仍面临相当大的挑战。Loyalka等人对中、美、印、俄四国STEM专业大学生的追踪调查发现,中国大学生在入学时批判性思维能力与他国相当,但在四年后出现显著下降。因此,仅依赖逻辑训练或知识灌输可能不足以支撑学生批判性思维的持续发展。 近年来,来自认知心理学和神经科学的研究为理解批判性思维提供了新视角。研究指出,批判性思维不仅依赖推理技能,还受到个体能否抑制系统1产生的直觉性优势反应、成功调动系统2进行逻辑加工的制约。例如,在物理学习中,学习者可能会因套用直觉或经验法则而作出错误判断。这些快速反应往往具有高度自动化特征,难以被意识到,因而需要抑制控制的参与才能启动反思性或批判性的思维加工。 本文基于双加工理论与抑制控制理论,指出物理学习中批判性思维的认知基础在于学习者能否识别并抑制直觉优势反应,转而调动系统2进行理性思考,这种抑制系统1和调动系统2的能力与执行功能中的抑制控制密切相关。本文在回顾批判性思维内涵的基础上,通过三个典型的物理学习案例揭示批判性思维在物理学习中的表现,进一步从理论和实证层面梳理了抑制控制与批判性思维之间的关联。在此基础上,本文还从抑制控制的视角对物理教学中培养学生的批判性思维提出了建议。
摘 要: 批判性思维是促使学生深度理解物理概念并提升其解决复杂问题能力的关键因素。然而,传统教学多聚焦于批判性思维相关的分析推理技能,较少关注批判性思维的认知基础。通过分析三个典型的物理学习案例,揭示批判性思维在物理学习中的表现特征。在此基础上,基于双加工理论和抑制控制理论,提出物理学习中批判性思维的认知基础在于学生能否识别并抑制直觉优势反应,转而调动系统2进行理性思考。接着,从认知行为与神经机制两个方面梳理抑制控制与批判性思维的实证证据。最后,从抑制控制训练融入物理课堂、提供支持性的物理课堂环境、抑制控制的自我监控与实践等方面提出对物理教学的启示。 批判性思维是21世纪人才应该具备的关键能力,尤其在科学教育领域,其对于学生形成理性判断、批判评估与创新表达的能力尤为重要。习近平总书记在中国科协第十次全国代表大会上的讲话中特别指出:“要更加重视人才自主培养,更加重视科学精神、创新能力、批判性思维的培养培育。”国内外科学课程亦高度强调批判性思维的重要性,如《普通高中物理课程标准(2017年版2020年修订)》指出:“……需要培养学生能够基于事实证据和科学推理对不同观点和结论提出质疑和批判,进行检验和修正,进而提出创新性见解的能力与品质”。 然而,我国青少年的批判性思维能力发展仍面临相当大的挑战。Loyalka等人对中、美、印、俄四国STEM专业大学生的追踪调查发现,中国大学生在入学时批判性思维能力与他国相当,但在四年后出现显著下降。因此,仅依赖逻辑训练或知识灌输可能不足以支撑学生批判性思维的持续发展。 近年来,来自认知心理学和神经科学的研究为理解批判性思维提供了新视角。研究指出,批判性思维不仅依赖推理技能,还受到个体能否抑制系统1产生的直觉性优势反应、成功调动系统2进行逻辑加工的制约。例如,在物理学习中,学习者可能会因套用直觉或经验法则而作出错误判断。这些快速反应往往具有高度自动化特征,难以被意识到,因而需要抑制控制的参与才能启动反思性或批判性的思维加工。 本文基于双加工理论与抑制控制理论,指出物理学习中批判性思维的认知基础在于学习者能否识别并抑制直觉优势反应,转而调动系统2进行理性思考,这种抑制系统1和调动系统2的能力与执行功能中的抑制控制密切相关。本文在回顾批判性思维内涵的基础上,通过三个典型的物理学习案例揭示批判性思维在物理学习中的表现,进一步从理论和实证层面梳理了抑制控制与批判性思维之间的关联。在此基础上,本文还从抑制控制的视角对物理教学中培养学生的批判性思维提出了建议。
考点 1:《普通高中物理课程标准(2017 年版 2020 年修订)》 应译为"Physics Curriculum Standards for Senior High Schools (2017 Edition, 2020 Revision)" 考点 2:STEM 专业大学生 应译为 "STEM major college students" 考点 3:追踪调查 应译为" longitudinal study /tracking survey” 考点4:高度自动化特征 应译为" highly automated characteristics"
3ca4a
学术论文
社会科学
97
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 近年来,在系列助农政策的支持下,我国特色林果的机械化生产技术得到迅速发展。然而,采收作为特色林果生产过程中劳动强度最大的环节,其机械化技术水平偏低已成为制约我国特色林果产业发展的主要问题。为全面了解我国特色林果机械化采收技术的发展现状和存在问题,系统梳理2018—2023年国内外在林果机械化采收领域的相关技术与装备研究,并进行深入分析。目前我国针对林果机械化采收技术的研究大部分停留在理论层面,落地用于实际生产的设备较少,且现有的机械采收设备大多需要人工辅助进行作业,功能相对单一且自动化程度较低。在此基础上提出我国特色林果采收设备应向自动化、智能化以及功能多样化等方向发展的思路,为进一步推动我国特色林果产业的机械化生产奠定基础。 近年来,桑葚、红枣、核桃等特色林果的种植面积生产过程中机械化程度逐渐增加,全过程自动化生产方式成为特色林果生产的发展趋势,其中机械化采收直接关系到特色林果产品的产出质量,是特色林果产业转变发展方式、节本增效、增强国际市场竞争力的重要途径之一[1],大力发展林果业机械化采收对促进我国特色林果产业发展具有重要意义。 林果机械化采收技术在国外已较为成熟,广泛应用于柑橘、樱桃、杏和红枣等林果的采收环节。然而,这些采收设备因其较强的针对性,难以适应我国果园的地形和多样化种植模式。相比之下,我国林果机械化采收技术起步较晚,技术层面尚有诸多不足。因此,研发适合我国林果园机械化作业的采收设备,对于推动我国特色林果产业迈向全程全面机械化生产具有重要意义。 本文基于现有的特色林果机械化采收技术,深入剖析我国林果机械化采收设备在行走、采摘和收获等作业环节中存在的问题,并针对这些问题提出切实可行的解决方案,以满足我国林果收获的实际需求,为我国特色林果机械化生产技术的持续发展提供参考。 1振动式采收 振动采收是我国果园使用最广泛的机械化采收方式之一,其工作原理为:通过振动器直接接触果树,使果树以一定的频率和振幅振动,带动果实产生惯性力,当惯性力大于果实的脱落阻力时,果实掉落,达到采摘果实的目的。振动采摘的工作原理决定了该方式的适用对象为果实较坚硬、对采摘质量要求不高的干果类、坚果类和部分鲜果类林果。 在研发振动采收设备之前,国内外学者通过构建果树模型并进行了深入的分析,研究了果树在受到振动时各部分的响应情况及不同振动位置对果实的损伤情况[2]。基于不同振动位置的采收效果,将林果振动式采收机分为树冠振动式、树枝振摇式(枝条)和树干振动式3种类型。 1.1树冠振动式采收 树冠振动式采收装置通常依靠指拍杆或拨棒插入果树冠层,使用动力装置驱动指拍杆以一定轨迹运动,以带动冠层枝条进行振动,从而达到落果的效果。在国外,树冠振动采收技术发展较为成熟,不仅实现了对一般林果在田间环境下的机械化采收,对于易受损伤的葡萄等浆果脱粒作业亦具备实现可能性。Caprara[3]和Pellenc[4]等通过动力装置驱动指拍杆击打葡萄藤实现葡萄采收,Ioan[5]利用偏心圆盘摇杆机构驱动尼龙弹性杆振摇葡萄藤实现酿酒葡萄采收。在我国,树冠振动技术的作业对象仍以对外观破损要求较低的红枣、核桃等林果为主。为了适应我国新疆红枣的矮化密植种植模式,付威等[6]设计了一种树冠振动式采收机———4ZZ-4自走式红枣收获机,梅松等[7]研究了通过拍打红枣树冠层枝条来实现红枣果实脱落。此外,树冠振动采收方式在核桃、油茶果、柑橘等果树的采收作业中应用广泛,王真真等[8]通过研究驱动拨杆转动来拍打核桃冠层,有效地降低了采收过程中装置对树枝的机械损伤;吴问天[9]、伍德林[10]和杜小强[11]等针对油茶果树的生长特点和果实脱落时的受力情况设计了适合油茶采收的树冠式振动采收装置;蒲应俊[12]根据柑橘树冠形态和果实分布特点设计了一种上、下两段式树冠振动装置,当上、下振动系统的振动频率分别为4.7Hz、4.1Hz时,其采收率和树冠损伤率分别为82.5%、3.9%。已有的树冠振动式采收机具有较高的采收效率和采净率,适合大面积种植的小型林果采收,且选择合适的振动力度和频率进行作业,对果树冠层枝条的损伤较小。树冠振动采收装置对红枣、枸杞等小型林果的采收效果较好[13]。
近年来,在系列助农政策的支持下,我国特色林果的机械化生产技术得到迅速发展。然而,采收作为特色林果生产过程中劳动强度最大的环节,其机械化技术水平偏低已成为制约我国特色林果产业发展的主要问题。为全面了解我国特色林果机械化采收技术的发展现状和存在问题,系统梳理2018—2023年国内外在林果机械化采收领域的相关技术与装备研究,并进行深入分析。目前我国针对林果机械化采收技术的研究大部分停留在理论层面,落地用于实际生产的设备较少,且现有的机械采收设备大多需要人工辅助进行作业,功能相对单一且自动化程度较低。在此基础上提出我国特色林果采收设备应向自动化、智能化以及功能多样化等方向发展的思路,为进一步推动我国特色林果产业的机械化生产奠定基础。 近年来,桑葚、红枣、核桃等特色林果的种植面积生产过程中机械化程度逐渐增加,全过程自动化生产方式成为特色林果生产的发展趋势,其中机械化采收直接关系到特色林果产品的产出质量,是特色林果产业转变发展方式、节本增效、增强国际市场竞争力的重要途径之一[1],大力发展林果业机械化采收对促进我国特色林果产业发展具有重要意义。 林果机械化采收技术在国外已较为成熟,广泛应用于柑橘、樱桃、杏和红枣等林果的采收环节。然而,这些采收设备因其较强的针对性,难以适应我国果园的地形和多样化种植模式。相比之下,我国林果机械化采收技术起步较晚,技术层面尚有诸多不足。因此,研发适合我国林果园机械化作业的采收设备,对于推动我国特色林果产业迈向全程全面机械化生产具有重要意义。 本文基于现有的特色林果机械化采收技术,深入剖析我国林果机械化采收设备在行走、采摘和收获等作业环节中存在的问题,并针对这些问题提出切实可行的解决方案,以满足我国林果收获的实际需求,为我国特色林果机械化生产技术的持续发展提供参考。 1振动式采收 振动采收是我国果园使用最广泛的机械化采收方式之一,其工作原理为:通过振动器直接接触果树,使果树以一定的频率和振幅振动,带动果实产生惯性力,当惯性力大于果实的脱落阻力时,果实掉落,达到采摘果实的目的。振动采摘的工作原理决定了该方式的适用对象为果实较坚硬、对采摘质量要求不高的干果类、坚果类和部分鲜果类林果。 在研发振动采收设备之前,国内外学者通过构建果树模型并进行了深入的分析,研究了果树在受到振动时各部分的响应情况及不同振动位置对果实的损伤情况[2]。基于不同振动位置的采收效果,将林果振动式采收机分为树冠振动式、树枝振摇式(枝条)和树干振动式3种类型。 1.1树冠振动式采收 树冠振动式采收装置通常依靠指拍杆或拨棒插入果树冠层,使用动力装置驱动指拍杆以一定轨迹运动,以带动冠层枝条进行振动,从而达到落果的效果。在国外,树冠振动采收技术发展较为成熟,不仅实现了对一般林果在田间环境下的机械化采收,对于易受损伤的葡萄等浆果脱粒作业亦具备实现可能性。Caprara[3]和Pellenc[4]等通过动力装置驱动指拍杆击打葡萄藤实现葡萄采收,Ioan[5]利用偏心圆盘摇杆机构驱动尼龙弹性杆振摇葡萄藤实现酿酒葡萄采收。在我国,树冠振动技术的作业对象仍以对外观破损要求较低的红枣、核桃等林果为主。为了适应我国新疆红枣的矮化密植种植模式,付威等[6]设计了一种树冠振动式采收机———4ZZ-4自走式红枣收获机,梅松等[7]研究了通过拍打红枣树冠层枝条来实现红枣果实脱落。此外,树冠振动采收方式在核桃、油茶果、柑橘等果树的采收作业中应用广泛,王真真等[8]通过研究驱动拨杆转动来拍打核桃冠层,有效地降低了采收过程中装置对树枝的机械损伤;吴问天[9]、伍德林[10]和杜小强[11]等针对油茶果树的生长特点和果实脱落时的受力情况设计了适合油茶采收的树冠式振动采收装置;蒲应俊[12]根据柑橘树冠形态和果实分布特点设计了一种上、下两段式树冠振动装置,当上、下振动系统的振动频率分别为4.7Hz、4.1Hz时,其采收率和树冠损伤率分别为82.5%、3.9%。已有的树冠振动式采收机具有较高的采收效率和采净率,适合大面积种植的小型林果采收,且选择合适的振动力度和频率进行作业,对果树冠层枝条的损伤较小。树冠振动采收装置对红枣、枸杞等小型林果的采收效果较好[13]。
考点 1:“国际市场竞争力” 推荐译为 “competitiveness in the international market” 考点 2:“干果类” 推荐译为 “dry fruits” 考点 3:“树冠振动式” 推荐译为 “canopy shaker type” 考点 4:“树枝振摇式” 推荐译为 “branch shaker type” 考点 5:“树干振动式” 推荐译为 “trunk shaker type” 考点 6:“指拍杆” 必须译为 “finger-beating rod” 考点7:“浆果脱粒作业”推荐译为“berry destemming operations”
3d0ab
学术论文
应用学科
173
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 二、物流行业未来发展趋势 2.1 物流运行保持良好增长 展望2025年,随着更加积极的财政政策和适度宽松的货币政策贯彻落实,政策“组合拳”将加快落地,社会物流需求有望得到提振。我国物流市场将保持温和增长态势,助力经济持续回升向好。这一趋势不仅得益于政策环境的优化,还受益于物流行业自身的结构调整和转型升级。 2.2 提质增效降本走向深入 物流降成本的出发点是服务实体经济,重点是通过提质增效降低供应链全链条的物流成本。受增长压力和降本要求影响,越来越多的工商企业、物流企业将逐步从流程优化、资源整合向更加深入的组织协同转变。在企业内部、企业之间、产业之间打破市场边界,加强与上下游、区域间联动融合,有望打开降本增效新空间,挖掘企业“第三利润源”,助力增强产业竞争力。 2.3 供应链升级加快提速 我国传统的运输、仓储业态已基本完成向现代物流服务的转型升级,下一阶段将加快向现代供应链转型升级。供应链是物流的高级形态,致力于通过生产组织方式的变革来创造新价值。随着产业高端化趋势显现,工商企业与物流企业将深化供应链融合创新,提供专业化、一体化、集成化的供应链解决方案,满足客户多样化需求,共同实现供应链价值创造。 三、物流行业未来发展趋势前景 3.1 韧性安全水平持续提升 我国出口货物结构发生根本性转变,从劳动密集型产品转向装备制造等中高端产品。同时,制造业“出海”将更加深入,逐步向“本地化生产、全球化流通”转变。这将推动现代物流加快从“跟随”战略转向“本地化”战略,加大国际物流市场布局和资源投入,通过与国际合作伙伴加强合作,共同构筑韧性安全的全球供应链服务体系,更好服务高水平对外开放。 3.2 数智化转型提档升级 数字经济与现代物流的融合发展走在各行业前列,我国物流领域基本具备了数字化能力。下一步,通过数字技术和物流组织的有机连接,实现更高水平的效能提升。将先进的数字技术与现代供应链的组织方式深度融合,大力发展平台经济,形成创新驱动的生产组织方式和运行模式,有望推动现代物流从信息化向智能化的跃迁,构建形成随需应变的数字供应链,开拓产业发展新赛道。 3.3 结构调整力度有所加大 随着国家大力提振消费,超大规模市场优势显现,消费物流的市场贡献将进一步增长。外部市场环境倒逼制造业加快向高端化升级,工业物流市场中的装备制造物流需求仍将保持韧性,支柱地位将进一步得到巩固。随着铁路货运深化市场改革,“铁路进码头”“白货上铁路”将成为趋势,运输结构持续得到优化。
二、物流行业未来发展趋势 2.1 物流运行保持良好增长 展望2025年,随着更加积极的财政政策和适度宽松的货币政策贯彻落实,政策“组合拳”将加快落地,社会物流需求有望得到提振。我国物流市场将保持温和增长态势,助力经济持续回升向好。这一趋势不仅得益于政策环境的优化,还受益于物流行业自身的结构调整和转型升级。 2.2 提质增效降本走向深入 物流降成本的出发点是服务实体经济,重点是通过提质增效降低供应链全链条的物流成本。受增长压力和降本要求影响,越来越多的工商企业、物流企业将逐步从流程优化、资源整合向更加深入的组织协同转变。在企业内部、企业之间、产业之间打破市场边界,加强与上下游、区域间联动融合,有望打开降本增效新空间,挖掘企业“第三利润源”,助力增强产业竞争力。 2.3 供应链升级加快提速 我国传统的运输、仓储业态已基本完成向现代物流服务的转型升级,下一阶段将加快向现代供应链转型升级。供应链是物流的高级形态,致力于通过生产组织方式的变革来创造新价值。随着产业高端化趋势显现,工商企业与物流企业将深化供应链融合创新,提供专业化、一体化、集成化的供应链解决方案,满足客户多样化需求,共同实现供应链价值创造。 三、物流行业未来发展趋势前景 3.1 韧性安全水平持续提升 我国出口货物结构发生根本性转变,从劳动密集型产品转向装备制造等中高端产品。同时,制造业“出海”将更加深入,逐步向“本地化生产、全球化流通”转变。这将推动现代物流加快从“跟随”战略转向“本地化”战略,加大国际物流市场布局和资源投入,通过与国际合作伙伴加强合作,共同构筑韧性安全的全球供应链服务体系,更好服务高水平对外开放。 3.2 数智化转型提档升级 数字经济与现代物流的融合发展走在各行业前列,我国物流领域基本具备了数字化能力。下一步,通过数字技术和物流组织的有机连接,实现更高水平的效能提升。将先进的数字技术与现代供应链的组织方式深度融合,大力发展平台经济,形成创新驱动的生产组织方式和运行模式,有望推动现代物流从信息化向智能化的跃迁,构建形成随需应变的数字供应链,开拓产业发展新赛道。 3.3 结构调整力度有所加大 随着国家大力提振消费,超大规模市场优势显现,消费物流的市场贡献将进一步增长。外部市场环境倒逼制造业加快向高端化升级,工业物流市场中的装备制造物流需求仍将保持韧性,支柱地位将进一步得到巩固。随着铁路货运深化市场改革,“铁路进码头”“白货上铁路”将成为趋势,运输结构持续得到优化。
考点1:“政策‘组合拳’”推荐译为“policy package”。 考点2:”高端化“不可简单译为“upgrading” 考点3:“业态”推荐译为“business format”。 考点4:“第三利润源“应该译为“the third profit source”是,一个特定的经济学概念。 考点5:”白货“应该译为“general cargo”,是铁路货运离你关于的专业术语,指包装规整、价值较高的工业制成品和消费品。
3e11b
学术论文
社会科学
37
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 移动、引拍、击球——测试场地上,一个四足机器人灵活地挥动机械臂,将飞来的羽毛球打回给它的人类对手。近日,瑞士苏黎世联邦理工学院研究团队在国际期刊《科学·机器人学》上发布了其研发的新型足式机器人系统(见图,瑞士苏黎世联邦理工学院供图)。这个特殊的“运动员”能够仅凭机载感知设备预测羽毛球飞行轨迹、调整自身位置并精确完成击打动作。“我们的成果展示了足式机器人如何执行复杂、动态、由感知驱动的任务,可以为整合机器人高速感知和全身协调能力提供新的思路。”该研究的第一作者、苏黎世联邦理工学院机器人系统实验室的马云涛博士在接受本报记者采访时表示。 羽毛球是世界上速度最快的球类运动之一,球速最高可达每小时400公里以上。运动员需要在极短的时间内捕捉羽毛球飞行轨迹,然后在一定范围内快速完成位移和击球动作。这项运动对预判、跑位和挥拍技术要求比较高,因而成为测试机器人动态感知与运动协调能力的理想项目。马云涛解释说,选择羽毛球作为实验对象,通过设置固定和移动等不同难度的击打目标,研究团队可以渐进式地检验和提升机器人性能。 要成为合格的“羽毛球选手”,机器人首先要做到“看得清”。现有商用机载相机的运动稳定性、可变焦距和信息处理能力等方面远逊于人眼,为实现有效的视觉追踪,研究团队开发了感知噪声模型,用以量化机器人抖动、旋转等运动状态对目标追踪的影响,使机器人能够主动适应动态模糊、目标遮挡等干扰。即便目标因高速运动或遮挡短暂消失,机器人仍可基于历史运动轨迹持续预测其位置。当羽毛球高速飞离视野中心时,机器人也会主动调整身体俯仰角度,将目标保持在相机视野内以优化追踪效果。 让机器人“动得准”是另一大挑战。传统运动机器人往往将移动与操作任务分离——底盘负责跑位,而机械臂负责操作。这种“各司其职”的设计使机器人难以应对复杂的动态环境。“我们通过一个基于强化学习的统一控制框架,将主动感知、移动和操作功能整合为一体。”该研究的共同作者、苏黎世联邦理工学院机械与加工工程系教授马科·胡特表示,这项技术可同步协调机器人周身18个关节的运动,通过判断来球的时间和距离,自主调整步态和击球方式。实验显示,在测试条件下,机器人可以在单个回合内与人类对手进行10次连续对打,且对于落在球场中心区域的球达到近100%的拦截成功率。 目前,机器人从发现对手击球到发出挥拍动作平均需要约0.35秒,其感知和反应能力仍有提升空间。研究团队计划通过集成更多传感器、融合多种传感模式并优化视觉算法等,进一步升级机器人性能。未来,这一成果有望走出球场,应用于更多需要快速响应和全身协调的复杂场景。“机器人学会了如何平衡感知稳定性与运动灵活性,这种视觉与运动之间的紧密耦合不仅对体育运动至关重要,对于灾难响应、人机协作等动态应用场景同样关键。这是迈向更具感知力与响应能力的移动机器人的重要一步。”马科·胡特表示。 “相较于轮式和履带式机器人,足式机器人具有更强的通用性,能够适应更广泛的应用场景,加之与人工智能技术的融合,将具备强大的感知和操作功能。”中国社会科学院中国式现代化研究院研究员李晓华表示,“随着人工智能与机器人本体技术的突破及深度融合,规模化生产促进生产成本显著下降,功能更强大、价格更低廉的足式机器人将在工业、休闲娱乐、居家生活、养老照护等领域获得广泛应用,发展成为对国民经济具有重要推动力和影响力的新兴产业。
移动、引拍、击球——测试场地上,一个四足机器人灵活地挥动机械臂,将飞来的羽毛球打回给它的人类对手。近日,瑞士苏黎世联邦理工学院研究团队在国际期刊《科学·机器人学》上发布了其研发的新型足式机器人系统(见图,瑞士苏黎世联邦理工学院供图)。这个特殊的“运动员”能够仅凭机载感知设备预测羽毛球飞行轨迹、调整自身位置并精确完成击打动作。“我们的成果展示了足式机器人如何执行复杂、动态、由感知驱动的任务,可以为整合机器人高速感知和全身协调能力提供新的思路。”该研究的第一作者、苏黎世联邦理工学院机器人系统实验室的马云涛博士在接受本报记者采访时表示。 羽毛球是世界上速度最快的球类运动之一,球速最高可达每小时400公里以上。运动员需要在极短的时间内捕捉羽毛球飞行轨迹,然后在一定范围内快速完成位移和击球动作。这项运动对预判、跑位和挥拍技术要求比较高,因而成为测试机器人动态感知与运动协调能力的理想项目。马云涛解释说,选择羽毛球作为实验对象,通过设置固定和移动等不同难度的击打目标,研究团队可以渐进式地检验和提升机器人性能。 要成为合格的“羽毛球选手”,机器人首先要做到“看得清”。现有商用机载相机的运动稳定性、可变焦距和信息处理能力等方面远逊于人眼,为实现有效的视觉追踪,研究团队开发了感知噪声模型,用以量化机器人抖动、旋转等运动状态对目标追踪的影响,使机器人能够主动适应动态模糊、目标遮挡等干扰。即便目标因高速运动或遮挡短暂消失,机器人仍可基于历史运动轨迹持续预测其位置。当羽毛球高速飞离视野中心时,机器人也会主动调整身体俯仰角度,将目标保持在相机视野内以优化追踪效果。 让机器人“动得准”是另一大挑战。传统运动机器人往往将移动与操作任务分离——底盘负责跑位,而机械臂负责操作。这种“各司其职”的设计使机器人难以应对复杂的动态环境。“我们通过一个基于强化学习的统一控制框架,将主动感知、移动和操作功能整合为一体。”该研究的共同作者、苏黎世联邦理工学院机械与加工工程系教授马科·胡特表示,这项技术可同步协调机器人周身18个关节的运动,通过判断来球的时间和距离,自主调整步态和击球方式。实验显示,在测试条件下,机器人可以在单个回合内与人类对手进行10次连续对打,且对于落在球场中心区域的球达到近100%的拦截成功率。 目前,机器人从发现对手击球到发出挥拍动作平均需要约0.35秒,其感知和反应能力仍有提升空间。研究团队计划通过集成更多传感器、融合多种传感模式并优化视觉算法等,进一步升级机器人性能。未来,这一成果有望走出球场,应用于更多需要快速响应和全身协调的复杂场景。“机器人学会了如何平衡感知稳定性与运动灵活性,这种视觉与运动之间的紧密耦合不仅对体育运动至关重要,对于灾难响应、人机协作等动态应用场景同样关键。这是迈向更具感知力与响应能力的移动机器人的重要一步。”马科·胡特表示。 “相较于轮式和履带式机器人,足式机器人具有更强的通用性,能够适应更广泛的应用场景,加之与人工智能技术的融合,将具备强大的感知和操作功能。”中国社会科学院中国式现代化研究院研究员李晓华表示,“随着人工智能与机器人本体技术的突破及深度融合,规模化生产促进生产成本显著下降,功能更强大、价格更低廉的足式机器人将在工业、休闲娱乐、居家生活、养老照护等领域获得广泛应用,发展成为对国民经济具有重要推动力和影响力的新兴产业。
考点 1:“四足机器人”应译为 "quadrupedal robot" 考点 2:“机载感知设备”应译为 "onboard perception system" 考点3:“足式机器人系统” 应译为 "legged robotic system" 考点4:“行轨迹 ”应译为 "flight trajectory" 考点5:“运动协调能力 ”应译为 "motor coordination ability" 考点6:“动态模糊”应译为 "motion blur" 考点7:历史运动轨迹 应译为 "historical motion trajectory" 考点8:俯仰角度 应译为 "pitch angle" 考点9:新兴产业 应译为 "emerging industry" 考点10:“黎世联邦理工学院”应译为“ETH Zurich” 考点11: “《科学·机器人学》”应译为“Science Robotics”
3e122
新闻资讯
新闻报道
85
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 与《社会学基本概念》中“代表与支配”合一的行政集团相关的是韦伯所谓的“自由代表制”类型。在《支配的类型》中,“自由代表制”与占有代表特权的身份型代表和约束型代表形成明显的反差。其差别的焦点在于,“自由代表认为表达自己实在的信念才与自己被推举为代表相符。这类自由代表在韦伯看来是“支配者”而不是被代表者的代理人。对韦伯来说,自由代表和代议机构的结合(而不是代表制本身)是西方历史的独特动力。这种结合面临两个方向的基本挑战,即直接面对选民的领袖制度和政党的科层化。为后代学者熟知的现代政治家的天职观恰恰只能出现在政治经营的两个典型困境的考验--民粹政治家与政党机构化--之后。政治天职被政治经营考验,体现了韦伯对正当性支配之说的局限性的认识。 根本而言,自由代表制积极地扩展了《社会学基本概念》当中“代表和支配合一的说法而不是与之构成矛盾。它与《政治作为天职》的核心论点是相互呼应的,都是对天职观两分的深化。两种意义上的“政治为业”相互刺激,是现代政治家的伦理出现的基本前提,这决定了韦伯眼中的现代政治家其实相当“晚出”,而且只能以“责任伦理”作为根本的规范共性。 在“高度统一和最严格的纪律”的结合方面,政党选举打破了乡绅名士和国会议员的限制,将政党权力放在了可以持续推进组织化的人身上。要平衡专业官僚,政治的支配意志需要更强大的民意动员。党工、总统和议会选举出的首相(凯撒制领袖)能对抗科层制形成的一种压倒生命力的理念,一旦职官再度变成专业官员控制的位置,也就是出现一战前德国式的行政官员变成望族的那种情形,那么民意领袖和政党机器合一的制度就会面临最大的挑战。而从首席政治家开始,政党选举越来越成为政治家的试金石。无论在议会制下还是在直接民选制下,胜利党领袖在理论上都表明政治家在行使支配者的力量,而不只是作为行政集团的首脑。这些促使领袖获得代议制选举胜利的政治经营,尤其是体现在盎格鲁-萨克逊传统中的机构化,构成了韦伯笔下的“机器”。 韦伯在19世纪前半期的美国和1868年之后的英国看见了借助政党机器抑制望族和议会的民选领袖。民意领袖借助机器化的政党机构战胜了望族和国会,这个过程与《政治作为天职》开始的时候君主借助行政官僚抑制贵族制有些形式上的类似。例如格莱斯通和迪斯莱利均需要一个隐在演说的幕后、以让领袖得到足够选民支持为已任的“党魁”。党魁接受一切针对其不择手段的“政治伦理”的指责包括接受以清除党魁的腐败为已任的党内候选者的指责。在韦伯看来,被上流鄙视是他们“志业性”的一部分。 政治家的责任围绕着权力展开,认识到权力是手段,而且要从这一手段而不是一般性的道德原则中找到一种伦理的力量,这是政治家的独有使命。将权力作为手段来享受是政治家产生内在支柱的前提,这一点和上文所说的国家作为特殊的政治手段而产生规范性要求的命题有内在的呼应。不过,政治家对自己用权力来追求什么的回答只能是基于信念。和《弗莱堡演讲》不同,韦伯在这个演说中并没有告诉我们政治家信仰的“目标”应该以什么样的价值尺度来衡量。然而,韦伯对什么样的人格会为运用权力负责任却有清楚的标准。 如果政治伦理的特殊性在于权力背后的责任感,那么韦伯只提供了责任感的人格保证:政治人格在于热情、责任心和判断力,这是韦伯关于志业政治家人格最著名的结论。只有具备政治人格,才能在正确的距离下看待权力是所有政治的手段这件事情,并将正当性的追求返回到政治家本身。 这样,韦伯在演讲开始时对国家的定义和在文末对政治家的期待之间形成了有意思的对照:在对国家的定义当中,我们只能够讨论国家动用什么手段去实现目标,而不能讨论国家可以实现什么目标。而在这里,政治家要对作为政治手段的权力负责,却没有关于权力的目标的价值(正义)尺度可用。韦伯赋予政治一种伦理色彩的前提是人格,而非政治目标的价值。 “机器”的譬喻在《政治作为天职》当中被韦伯给予了政党机器,而不是行政官僚。党内那些“自觉”放弃自身信念,或是在思想上尽量简陋的追随者才最典型地代表了机器。领袖出现的必然条件--或者说代价--就是机器。在《经济与社会》的写作体系当中,机器的譬喻只在《支配社会学》中大量地出现,在《社会学基本概念》和《支配的类型》这两篇以概念界说为重的文字中出现得很少,甚至没有韦伯在《政治共同体》篇中使用得多。 领袖事实上能做到的,不取决于他自己,而是追随者行动的动机”。这些动机并不会和家产制支配之下有什么变化,都是“报复欲、权利欲、掠夺欲和俸禄欲”,而政治伦理不过是用再真诚不过的政治家理想让这些欲望在道德上正当化而已。信仰的英雄是消逝得最快的英雄,这就形成了一种悖论:如果目标的实现不取决于政治家自己的理想,而是追随者的动机,而追随政治家的人本质上是放弃或者剥脱信念的,那么所谓责任感,就首先是对无信念的政治负责任的伦理。“无情”的支配不来自于本不属于支配者的专业官员,而来自支配者自己。政治家“会意识到在这些悙谬的压力下所发生的改变,而且这改变是由他自己来负责任的”。 天职观的两分以政治家区分信念和魔鬼般的世界为前提,浸透了强烈的新教政治的风格。责任伦理是对政治没有伦理的鲜明回应。政治的经营是政治家无法摆脱的魔鬼般的力量,政治家的责任伦理的力量是始终和魔鬼相伴的力量,一种人世禁欲主义的体现。浮士德笔下和灵魂同在的魔鬼在韦伯看来才是政治家的最好写照。
与《社会学基本概念》中“代表与支配”合一的行政集团相关的是韦伯所谓的“自由代表制”类型。在《支配的类型》中,“自由代表制”与占有代表特权的身份型代表和约束型代表形成明显的反差。其差别的焦点在于,“自由代表认为表达自己实在的信念才与自己被推举为代表相符。这类自由代表在韦伯看来是“支配者”而不是被代表者的代理人。对韦伯来说,自由代表和代议机构的结合(而不是代表制本身)是西方历史的独特动力。这种结合面临两个方向的基本挑战,即直接面对选民的领袖制度和政党的科层化。为后代学者熟知的现代政治家的天职观恰恰只能出现在政治经营的两个典型困境的考验--民粹政治家与政党机构化--之后。政治天职被政治经营考验,体现了韦伯对正当性支配之说的局限性的认识。 根本而言,自由代表制积极地扩展了《社会学基本概念》当中“代表和支配合一的说法而不是与之构成矛盾。它与《政治作为天职》的核心论点是相互呼应的,都是对天职观两分的深化。两种意义上的“政治为业”相互刺激,是现代政治家的伦理出现的基本前提,这决定了韦伯眼中的现代政治家其实相当“晚出”,而且只能以“责任伦理”作为根本的规范共性。 在“高度统一和最严格的纪律”的结合方面,政党选举打破了乡绅名士和国会议员的限制,将政党权力放在了可以持续推进组织化的人身上。要平衡专业官僚,政治的支配意志需要更强大的民意动员。党工、总统和议会选举出的首相(凯撒制领袖)能对抗科层制形成的一种压倒生命力的理念,一旦职官再度变成专业官员控制的位置,也就是出现一战前德国式的行政官员变成望族的那种情形,那么民意领袖和政党机器合一的制度就会面临最大的挑战。而从首席政治家开始,政党选举越来越成为政治家的试金石。无论在议会制下还是在直接民选制下,胜利党领袖在理论上都表明政治家在行使支配者的力量,而不只是作为行政集团的首脑。这些促使领袖获得代议制选举胜利的政治经营,尤其是体现在盎格鲁-萨克逊传统中的机构化,构成了韦伯笔下的“机器”。 韦伯在19世纪前半期的美国和1868年之后的英国看见了借助政党机器抑制望族和议会的民选领袖。民意领袖借助机器化的政党机构战胜了望族和国会,这个过程与《政治作为天职》开始的时候君主借助行政官僚抑制贵族制有些形式上的类似。例如格莱斯通和迪斯莱利均需要一个隐在演说的幕后、以让领袖得到足够选民支持为已任的“党魁”。党魁接受一切针对其不择手段的“政治伦理”的指责包括接受以清除党魁的腐败为已任的党内候选者的指责。在韦伯看来,被上流鄙视是他们“志业性”的一部分。 政治家的责任围绕着权力展开,认识到权力是手段,而且要从这一手段而不是一般性的道德原则中找到一种伦理的力量,这是政治家的独有使命。将权力作为手段来享受是政治家产生内在支柱的前提,这一点和上文所说的国家作为特殊的政治手段而产生规范性要求的命题有内在的呼应。不过,政治家对自己用权力来追求什么的回答只能是基于信念。和《弗莱堡演讲》不同,韦伯在这个演说中并没有告诉我们政治家信仰的“目标”应该以什么样的价值尺度来衡量。然而,韦伯对什么样的人格会为运用权力负责任却有清楚的标准。 如果政治伦理的特殊性在于权力背后的责任感,那么韦伯只提供了责任感的人格保证:政治人格在于热情、责任心和判断力,这是韦伯关于志业政治家人格最著名的结论。只有具备政治人格,才能在正确的距离下看待权力是所有政治的手段这件事情,并将正当性的追求返回到政治家本身。 这样,韦伯在演讲开始时对国家的定义和在文末对政治家的期待之间形成了有意思的对照:在对国家的定义当中,我们只能够讨论国家动用什么手段去实现目标,而不能讨论国家可以实现什么目标。而在这里,政治家要对作为政治手段的权力负责,却没有关于权力的目标的价值(正义)尺度可用。韦伯赋予政治一种伦理色彩的前提是人格,而非政治目标的价值。 “机器”的譬喻在《政治作为天职》当中被韦伯给予了政党机器,而不是行政官僚。党内那些“自觉”放弃自身信念,或是在思想上尽量简陋的追随者才最典型地代表了机器。领袖出现的必然条件--或者说代价--就是机器。在《经济与社会》的写作体系当中,机器的譬喻只在《支配社会学》中大量地出现,在《社会学基本概念》和《支配的类型》这两篇以概念界说为重的文字中出现得很少,甚至没有韦伯在《政治共同体》篇中使用得多。 领袖事实上能做到的,不取决于他自己,而是追随者行动的动机”。这些动机并不会和家产制支配之下有什么变化,都是“报复欲、权利欲、掠夺欲和俸禄欲”,而政治伦理不过是用再真诚不过的政治家理想让这些欲望在道德上正当化而已。信仰的英雄是消逝得最快的英雄,这就形成了一种悖论:如果目标的实现不取决于政治家自己的理想,而是追随者的动机,而追随政治家的人本质上是放弃或者剥脱信念的,那么所谓责任感,就首先是对无信念的政治负责任的伦理。“无情”的支配不来自于本不属于支配者的专业官员,而来自支配者自己。政治家“会意识到在这些悙谬的压力下所发生的改变,而且这改变是由他自己来负责任的”。 天职观的两分以政治家区分信念和魔鬼般的世界为前提,浸透了强烈的新教政治的风格。责任伦理是对政治没有伦理的鲜明回应。政治的经营是政治家无法摆脱的魔鬼般的力量,政治家的责任伦理的力量是始终和魔鬼相伴的力量,一种人世禁欲主义的体现。浮士德笔下和灵魂同在的魔鬼在韦伯看来才是政治家的最好写照。
考点1. “支配”推荐译为“Domination” 考点2. "支配者”推荐译为“Ruler; Dominator” 考点3. “责任伦理”推荐译为“Ethic of Responsibility” 考点4. “自由代表制”推荐译为“Free Representation” 考点5. “乡绅名士”推荐译为“Notables” 考点6. “凯撒制领袖, 民意领袖”推荐译为“Plebiscitary Leadership; Caesarist Leader” 考点7. “机器”推荐译为“(Party) Machine” 考点8. “自觉放弃自身信念”推荐译为“Disenchantment of the soul; ” 考点9. “入世禁欲主义”推荐译为“Inner-worldly Asceticism”, 考点10. “政治经营”推荐译为“the operation of politics (as an enterprise) 考点11. ”行政集团”建议译为“administrative group; administrative corps” 考点12. “民粹政治家”建议译为“demagogue” 考点13. “政治人格”建议译为“political personality” 考点14. “志业性”建议译为“vocational character” 考点15.“天职, 志业”推荐译为“Vocation; Calling”
3fffd
学术论文
社会科学
123
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 具身智能研究范畴已从早期的感知 — 反应式系统,发展至当代融合非平稳动力学建模与神经符号混合推理的跨学科前沿。当前,研究者借鉴形态计算、感官运动偶合与主动推理等理论框架,试图构建具有自组织与自适应能力的物理智能体。形态计算关注利用机器人本体的物理属性以减轻控制计算负担,而感官运动偶合则强调感知输入与运动输出的共同涌现机制,为多自由度系统的柔性操控提供新思路。主动推理以自由能原理为基础,通过最小化对环境模型的不确定性,达成对动态场景的自发贝叶斯更新。 在具身智能的仿真与实证研究中,可微分仿真平台已成为关键工具。诸如可微分光照模拟、可微分流体动力学与可微分多体动力学的联合仿真,可通过梯度反向传播实现对物理参数的端到端训练。研究团队利用可微分仿真器与脉冲神经网络相结合,探索神经形态硬件在实时高频反馈任务中的可行性;并引入共轭梯度解算器对接全向移动底盘与仿生手爪的协同运动规划,以兼顾能量效率与轨迹精度。 从算法层面看,近年兴起的图注意力网络与拓扑流形学习方法,为刻画高维状态空间中的结构化约束提供了有效途径。通过构建基于图卷积的世界模型,研究者能够在多主体协作场景中,利用张量分解与谱聚类方法对交互关系进行稀疏编码;进一步结合基于变分贝叶斯的群体推理,实现对智能体集群行为的宏观策略优化。此外,混合整数规划与深度确定性策略梯度的混合框架,已被用于解决高维离散动作与连续控制一体化决策问题。 在生物启发层面,仿生柔体机器人与可重构结构的结合日趋深入。研究者通过脉冲编码的软体流变元件,实现对触觉 — 形变反馈的实时重构,并利用稀疏自编码器对触觉时序信号进行特征抽取。该方案在复杂管道导航与狭窄空间抓取任务中,展现出超越刚体机器人显著的鲁棒性与柔性适应能力。与此同时,基于仿生学的非线性共振控制技术,正被用于增强软体机构在高速振动环境中的稳定性。 具身智能中的层次化控制方法亦不断演进。自顶向下的层次化部分可由符号规划器输出高层目标节点序列,随后由基于混沌系统的低层连续控制器进行实时执行;而自底向上的子任务发现,则运用面向补偿学习的元强化学习算法,在不确定性环境中自主生成行为原型。这种双向耦合的控制范式,结合了形式化方法与统计学习的优势,能够在开放世界任务中快速收敛至次优或近似最优解。 在多模态感知方面,跨域时空对齐技术正在广泛应用。通过谱域散度最小化与流形正则化,研究者实现了视觉 — 激光雷达 — 红外成像等异构数据的统一表征;并借助基于能量模型的对抗学习,对传感器故障或数据稀疏情况下的鲁棒性进行系统评估。更进一步,融合自监督的周期一致性损失与几何约束正则化,能够在无标签环境下,自动生成高质量的语义地图。 工业界亦在快速跟进:面向智造车间的具身智能系统,已逐步从原型验证转向小规模部署。若干企业采用神经符号混合推理架构,结合可微分示教与在线迁移学习,实现了机器手臂在异构工件 — 不同表面材质 — 多工序装配中的自适应夹持与路线规划。与此同时,针对动态协作机器人与人类工人的交互安全,研究机构提出了基于约束优化的实时安全壳算法,以确保协作空间内的碰撞避免与轨迹协同。 开放研究挑战方面,首当其冲的是多尺度动力学的统一表征。现有模型多针对有限自由度或刚体假设,却难以兼顾非线性耦合、柔体变形与流体 - 固体交互等多物理场耦合问题。为弥补此短板,一些前沿研究将稀疏标量场分解与基于图的多物理仿真融合,通过自适应网格细化与基于高阶有限元的本构模型,对复杂拓扑变化过程进行可微化表达。该方法在预测软体水下推进机理与仿真生物诱导流场中取得了初步成果。 其次,具身智能安全与伦理审查尚未形成系统方法论。随著智能体具备更高自主决策能力,如何在决策流程中引入符合社会规范与法律法规的价值约束,成为亟待解决的跨学科课题。研究者正在探索基于博弈论的伦理多目标优化,将道德权衡映射为约束优化问题;并计划在仿真预演环节,引入虚拟伦理审计器,以量化智能体在不同场景下的伦理风险暴露。 面向未来,异构计算架构与硬 — 软件协同设计是推动具身智能实用化的关键。除神经形态芯片外,多核时序处理器、高性能 FPGA 与可重构 ASIC 的融合,能够为高维感知和闭环反馈提供低延迟与确定性性能;而软硬件协同优化框架,将围绕调度算法、功耗模型与实时操作系统内核层面进行深入耦合,为现场可部署的具身智能系统奠定基础。 总之,具身智能正处于从理论化探究向工程化应用的临界阶段。跨学科融合、仿真 — 现实闭环与安全伦理共建,将成为未来研究的三大主轴。唯有在动力学本质、认知推理与系统工程层面形成闭环迭代,方能真正迈入 “类人自主” 与 “自主可解释” 并重的物理智能新时代。
具身智能研究范畴已从早期的感知 — 反应式系统,发展至当代融合非平稳动力学建模与神经符号混合推理的跨学科前沿。当前,研究者借鉴形态计算、感官运动偶合与主动推理等理论框架,试图构建具有自组织与自适应能力的物理智能体。形态计算关注利用机器人本体的物理属性以减轻控制计算负担,而感官运动偶合则强调感知输入与运动输出的共同涌现机制,为多自由度系统的柔性操控提供新思路。主动推理以自由能原理为基础,通过最小化对环境模型的不确定性,达成对动态场景的自发贝叶斯更新。 在具身智能的仿真与实证研究中,可微分仿真平台已成为关键工具。诸如可微分光照模拟、可微分流体动力学与可微分多体动力学的联合仿真,可通过梯度反向传播实现对物理参数的端到端训练。研究团队利用可微分仿真器与脉冲神经网络相结合,探索神经形态硬件在实时高频反馈任务中的可行性;并引入共轭梯度解算器对接全向移动底盘与仿生手爪的协同运动规划,以兼顾能量效率与轨迹精度。 从算法层面看,近年兴起的图注意力网络与拓扑流形学习方法,为刻画高维状态空间中的结构化约束提供了有效途径。通过构建基于图卷积的世界模型,研究者能够在多主体协作场景中,利用张量分解与谱聚类方法对交互关系进行稀疏编码;进一步结合基于变分贝叶斯的群体推理,实现对智能体集群行为的宏观策略优化。此外,混合整数规划与深度确定性策略梯度的混合框架,已被用于解决高维离散动作与连续控制一体化决策问题。 在生物启发层面,仿生柔体机器人与可重构结构的结合日趋深入。研究者通过脉冲编码的软体流变元件,实现对触觉 — 形变反馈的实时重构,并利用稀疏自编码器对触觉时序信号进行特征抽取。该方案在复杂管道导航与狭窄空间抓取任务中,展现出超越刚体机器人显著的鲁棒性与柔性适应能力。与此同时,基于仿生学的非线性共振控制技术,正被用于增强软体机构在高速振动环境中的稳定性。 具身智能中的层次化控制方法亦不断演进。自顶向下的层次化部分可由符号规划器输出高层目标节点序列,随后由基于混沌系统的低层连续控制器进行实时执行;而自底向上的子任务发现,则运用面向补偿学习的元强化学习算法,在不确定性环境中自主生成行为原型。这种双向耦合的控制范式,结合了形式化方法与统计学习的优势,能够在开放世界任务中快速收敛至次优或近似最优解。 在多模态感知方面,跨域时空对齐技术正在广泛应用。通过谱域散度最小化与流形正则化,研究者实现了视觉 — 激光雷达 — 红外成像等异构数据的统一表征;并借助基于能量模型的对抗学习,对传感器故障或数据稀疏情况下的鲁棒性进行系统评估。更进一步,融合自监督的周期一致性损失与几何约束正则化,能够在无标签环境下,自动生成高质量的语义地图。 工业界亦在快速跟进:面向智造车间的具身智能系统,已逐步从原型验证转向小规模部署。若干企业采用神经符号混合推理架构,结合可微分示教与在线迁移学习,实现了机器手臂在异构工件 — 不同表面材质 — 多工序装配中的自适应夹持与路线规划。与此同时,针对动态协作机器人与人类工人的交互安全,研究机构提出了基于约束优化的实时安全壳算法,以确保协作空间内的碰撞避免与轨迹协同。 开放研究挑战方面,首当其冲的是多尺度动力学的统一表征。现有模型多针对有限自由度或刚体假设,却难以兼顾非线性耦合、柔体变形与流体 - 固体交互等多物理场耦合问题。为弥补此短板,一些前沿研究将稀疏标量场分解与基于图的多物理仿真融合,通过自适应网格细化与基于高阶有限元的本构模型,对复杂拓扑变化过程进行可微化表达。该方法在预测软体水下推进机理与仿真生物诱导流场中取得了初步成果。 其次,具身智能安全与伦理审查尚未形成系统方法论。随著智能体具备更高自主决策能力,如何在决策流程中引入符合社会规范与法律法规的价值约束,成为亟待解决的跨学科课题。研究者正在探索基于博弈论的伦理多目标优化,将道德权衡映射为约束优化问题;并计划在仿真预演环节,引入虚拟伦理审计器,以量化智能体在不同场景下的伦理风险暴露。 面向未来,异构计算架构与硬 — 软件协同设计是推动具身智能实用化的关键。除神经形态芯片外,多核时序处理器、高性能 FPGA 与可重构 ASIC 的融合,能够为高维感知和闭环反馈提供低延迟与确定性性能;而软硬件协同优化框架,将围绕调度算法、功耗模型与实时操作系统内核层面进行深入耦合,为现场可部署的具身智能系统奠定基础。 总之,具身智能正处于从理论化探究向工程化应用的临界阶段。跨学科融合、仿真 — 现实闭环与安全伦理共建,将成为未来研究的三大主轴。唯有在动力学本质、认知推理与系统工程层面形成闭环迭代,方能真正迈入 “类人自主” 与 “自主可解释” 并重的物理智能新时代。
考点1:“梯度反向传播”推荐译为“back propagation” 考点2:“软体流变元件”推荐译为“ rheological components” 考点3:“语义地图”推荐译为“semantic map” 考点4:“感官运动耦合”推荐译为“sensory-motor coupling”
4096a
学术论文
应用学科
168
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Insert Table 1 As shown in Table 1, diverse theoretical perspectives and methodological approaches have emerged in relation to the study of influencers. Follower and human influencer perspectives are common (Franke et al. 2023). Recently, the emergence of virtual influencers has provided brands with new ways to communicate with consumers (Ramadan et al. 2021). This development offers a fresh perspective on the role of influencers, particularly concerning the role of virtual influencers (e.g., Lou et al. 2023; Mouritzen et al. 2023) within traditionally human-focused social media ecosystems. Virtual influencers are computer-generated, animated characters that behave like real humans, and are usually controlled by firms or agencies (Hofeditz et al. 2022). They are becoming increasingly realistic, narrowing the gap between humans and virtual agents (Sands et al. 2022). They tend to gather a large number of followers (Thomas and Fowler 2020) as they offer consumers unique social interactions that bridge the real and imaginary worlds (Arsenyan and Mirowska 2021). Virtual influencers represent the direction in which brands embed artificial intelligence (AI) in their communications (Ahn et al., 2022), enabling new forms of interaction between consumers and AI-powered agents (Huh et al. 2023). Virtual influencers can offer brands a number of advantages over human influencers, such as content control, flexibility (Miao et al. 2021), and enhanced perceived ad novelty (Franke et al. 2023). Therefore, it is not surprising that brands leverage the unique relationship between virtual influencers and followers. However, our current 7 knowledge of this phenomenon is limited, because virtual influencers integrate both human and virtual features. To uncover the nature of the virtual influencer-follower relationship, there is a need to understand the identity of virtual influencer, as it defines their relationships with their followers (Mrad et al. 2022). To attest to their identity, virtual influencers rely on perceived anthropomorphism (Ahn et al. 2022). Anthropomorphism is an inductive inference process by which consumers attribute human-like characteristics, intentions, or mental states to nonhuman agents (Waytz et al. 2010). It can lower reactance towards virtual agents (Fernandes and Oliveira 2021), make interactions more effective,and influence consumer trust in promotional content (Touré-Tillery and McGill 2015). A growing number of researchers have explored the influence of anthropomorphism on consumer interactions with AI agents, particularly chatbots (e.g., Crolic et al. 2022). Anthropomorphism enhances consumers’ perceptions of chatbots’ social presence and emotional connections to firms (Araujo 2018). A highly anthropomorphic chatbot can prompt consumers’ perceptions of a human actor and satisfy their human interaction needs (Sheehan et al. 2020). However, the nature of interactions with chatbots differs from that with virtual influencers, and the findings from this literature may not apply to virtual influencers. Companies use chatbots to engage with consumers in conversational settings (Youn and Jin 2021). Their relationships with consumers are two-sided, short-lived, goal-oriented, and functional. Consumers engage in conversations with robots to perform tasks. Nonetheless, the relationship between virtual influencers and consumers is different, as it involves repeated interactions, leading to strong bonds and relational patterns (Mrad et al. 2022). Knowledge of the relationship with the audience is crucial for understanding the impact of virtual influencers on consumers. To date, researchers have addressed the relationship between human influencers and followers through parasocial relationship lenses, as discussed in the next section. A parasocial relationshipis readily applicable to human influencers (Labrecque 2014) but it is even more relevant in the case of virtual influencers, as they are non-sentient and unable to build bonds by themselves.However, consumers can develop an attachment to virtual influencers that resemble human interaction with the help of the human-like identity that virtual influencers present. Parasocial relationship theory Parasocial relationships are one-sided relationships that humans form with media figures or online characters such as social media influencers (Lou et al. 2023). Similarly, relationships between humans and virtual influencers fall within the parasocial relationship category. A sense of connection, relational intimacy, and emotions may develop even when the ‘subject’ is unaware of their existence. By definition, parasocial relationships are primarily one-sided because relational knowledge and interest remain unreciprocated (Escalas and Bettman 2017). Although there is a friendly atmosphere between the media persona and their audience (Horton and Wohl 1956), the media persona does not possess the same relational knowledge as the individuals in their audience. Parasocial relationships represent unreciprocated bonds driven by feelings of intimacy, kinship, and attachment. Traditionally, researchers have identified these dynamics in TV and radio contexts (Horton and Wohl 1956); however, they are also deemed relevant in various other settings, including social media, where influencers reveal life events and everyday moments that create relationship intimacy and perceived reciprocity among their followers (Lou 2021). Researchers have extended the notion of parasocial relationships to avatars, robots (Whang and Im 2021) and brands in general (Labrecque 2014). The frequency of engagement, including even low-level interactions, such as seeing others, helps in relationship development (Heide and Miner 1992). This is relevant in both real-life and online settings. Marketers draw on trust in parasocial relationships, for instance, in celebrities endorsing products (Chung and Cho 2017), which is well-received by consumers and increases their purchase intention to endorse products (Lou 2021). Nonetheless, developing a parasocial relationship also depends on perceived authenticity and credibility (Whang and Im 2021), including an organic fit between the influencer and the endorsed products/services. Perceived similarities strengthen relationships: humans are naturally drawn to others who have similar traits and characteristics, even if these similarities only exist at the level of perception and even if the subject of attraction is non-human (Whang and Im 2021). However, a non-human entity is typically perceived as genuine when it is anthropomorphized and human-like (Whang and Im 2021). Perceived mutuality is stronger with human-like non-humans (Whang and Im 2021), and consumers associate higher levels of consciousness and responsibility with them, as opposed to less human-like non-humans (Aggarwal and McGill 2007). Both the study of parasocial relationships and social comparison theory are concerned with relationships, social interactions, and self-worth. Parasocial relationships influence self-evaluation and perceived self-worth. For instance, when people compare themselves unfavorably to celebrity figures, their self-esteem may decrease (Staniewski and Awruk 2022). The existing literature in this area lacks an in-depth understanding of how anthropomorphization influences parasocial relationships and social comparisons between virtual influencers and their followers. The following section delves into the relevant aspects of social comparison theory. Social comparison theory Social comparison processes can influence identification; thus, followers may assess their position within a group of fans. This may be a human or non-human fan group; in either case, the parasocial relationships within the follower group can contribute to an individual’s social identity (Aw and Chuah 2021). Social comparison is a widely shared human trait that involves comparing ourselves to others and is an important element of human life. This theory explains how individuals compare themselves to others and gain or lose self-esteem, depending on evaluations of their abilities, achievements, beliefs, attitudes, and appearances compared to those of their peers (Festinger 1954). Advertising scholarship shows that social comparisons to idealized celebrities can lower one’s self-esteem related to physical attractiveness and financial success (Gulas and McKeage 2000), and increase the sharing of materialistic values enacted through imitation, such as the purchase of products that celebrities possess (Chan and Prendergast 2008). Goethals (1986) views social comparison theory as a natural extension of researching how people evaluate themselves, approach their reference groups, and the comparative functions reference groups provide, including pressure towards conformity. Suls and Wheeler (2013) identified influencing factors in social comparison theory, such as individuals’ adaptiveness, egocentrism, tendency for projecting, approaches to belief formation, and whether individuals critically revisit or review beliefs. Social comparisons may lead to both positive (e.g., feeling motivated to improve) and negative emotions. For instance, envy towards those perceived as more attractive may lead to undervaluing them in other areas, such as professional skills (Bower 2001).Incorporating social comparison theory into our theorization opens up versatile possibilities. Comparisons are inherent in relationships with others and in one’s own relationships with oneself. Researchers have used it to study burnout (Buunk and Schaufeli 2018) and deception regarding consumption (Argo et al. 2006). This theory identifies multiple ways of making comparisons. Comparisons can range from extremely negative (upward comparisons) to highly positive valence (downward comparisons) (Goethals 1986). Examples of negativity relate to a lack of selfappreciation, depressive feelings, distress, and offending others, whereas positivity is typically about feeling more motivated and grateful as a result of social comparisons. Despite their practical relevance, little is known about social comparisons that occur between humans in relation to virtual influencers. Hence, this study addresses the following research question: How does the virtual influencer phenomenon influence followers’ online social experiences through social comparisons and parasocial relationships? This study sought to uncover how social comparisons and parasocial relationships manifest in the virtual influencer phenomenon. It aims to further understand how individuals perceive themselves in comparison to virtual influencers and how they form emotional relationships with them. Utilizing a theory of knitting approach following De Pelecijn et al. (2023), synergies have been identified between the key theoretical streams of anthropomorphization, parasocial relationships, and social comparison theory:relevance, little is known about social comparisons that occur between humans in relation to virtual influencers. Hence, this study addresses the following research question: How does the virtual influencer phenomenon influence followers’ online social experiences through social comparisons and parasocial relationships? This study sought to uncover how social comparisons and parasocial relationships manifest in the virtual influencer phenomenon. It aims to further understand how individuals perceive themselves in comparison to virtual influencers and how they form emotional relationships with them. Utilizing a theory of knitting approach following De Pelecijn et al. (2023), synergies have been identified between the key theoretical streams of anthropomorphization, parasocial relationships, and social comparison theory:
Insert Table 1 As shown in Table 1, diverse theoretical perspectives and methodological approaches have emerged in relation to the study of influencers. Follower and human influencer perspectives are common (Franke et al. 2023). Recently, the emergence of virtual influencers has provided brands with new ways to communicate with consumers (Ramadan et al. 2021). This development offers a fresh perspective on the role of influencers, particularly concerning the role of virtual influencers (e.g., Lou et al. 2023; Mouritzen et al. 2023) within traditionally human-focused social media ecosystems. Virtual influencers are computer-generated, animated characters that behave like real humans, and are usually controlled by firms or agencies (Hofeditz et al. 2022). They are becoming increasingly realistic, narrowing the gap between humans and virtual agents (Sands et al. 2022). They tend to gather a large number of followers (Thomas and Fowler 2020) as they offer consumers unique social interactions that bridge the real and imaginary worlds (Arsenyan and Mirowska 2021). Virtual influencers represent the direction in which brands embed artificial intelligence (AI) in their communications (Ahn et al., 2022), enabling new forms of interaction between consumers and AI-powered agents (Huh et al. 2023). Virtual influencers can offer brands a number of advantages over human influencers, such as content control, flexibility (Miao et al. 2021), and enhanced perceived ad novelty (Franke et al. 2023). Therefore, it is not surprising that brands leverage the unique relationship between virtual influencers and followers. However, our current 7 knowledge of this phenomenon is limited, because virtual influencers integrate both human and virtual features. To uncover the nature of the virtual influencer-follower relationship, there is a need to understand the identity of virtual influencer, as it defines their relationships with their followers (Mrad et al. 2022). To attest to their identity, virtual influencers rely on perceived anthropomorphism (Ahn et al. 2022). Anthropomorphism is an inductive inference process by which consumers attribute human-like characteristics, intentions, or mental states to nonhuman agents (Waytz et al. 2010). It can lower reactance towards virtual agents (Fernandes and Oliveira 2021), make interactions more effective,and influence consumer trust in promotional content (Touré-Tillery and McGill 2015). A growing number of researchers have explored the influence of anthropomorphism on consumer interactions with AI agents, particularly chatbots (e.g., Crolic et al. 2022). Anthropomorphism enhances consumers’ perceptions of chatbots’ social presence and emotional connections to firms (Araujo 2018). A highly anthropomorphic chatbot can prompt consumers’ perceptions of a human actor and satisfy their human interaction needs (Sheehan et al. 2020). However, the nature of interactions with chatbots differs from that with virtual influencers, and the findings from this literature may not apply to virtual influencers. Companies use chatbots to engage with consumers in conversational settings (Youn and Jin 2021). Their relationships with consumers are two-sided, short-lived, goal-oriented, and functional. Consumers engage in conversations with robots to perform tasks. Nonetheless, the relationship between virtual influencers and consumers is different, as it involves repeated interactions, leading to strong bonds and relational patterns (Mrad et al. 2022). Knowledge of the relationship with the audience is crucial for understanding the impact of virtual influencers on consumers. To date, researchers have addressed the relationship between human influencers and followers through parasocial relationship lenses, as discussed in the next section. A parasocial relationshipis readily applicable to human influencers (Labrecque 2014) but it is even more relevant in the case of virtual influencers, as they are non-sentient and unable to build bonds by themselves.However, consumers can develop an attachment to virtual influencers that resemble human interaction with the help of the human-like identity that virtual influencers present. Parasocial relationship theory Parasocial relationships are one-sided relationships that humans form with media figures or online characters such as social media influencers (Lou et al. 2023). Similarly, relationships between humans and virtual influencers fall within the parasocial relationship category. A sense of connection, relational intimacy, and emotions may develop even when the ‘subject’ is unaware of their existence. By definition, parasocial relationships are primarily one-sided because relational knowledge and interest remain unreciprocated (Escalas and Bettman 2017). Although there is a friendly atmosphere between the media persona and their audience (Horton and Wohl 1956), the media persona does not possess the same relational knowledge as the individuals in their audience. Parasocial relationships represent unreciprocated bonds driven by feelings of intimacy, kinship, and attachment. Traditionally, researchers have identified these dynamics in TV and radio contexts (Horton and Wohl 1956); however, they are also deemed relevant in various other settings, including social media, where influencers reveal life events and everyday moments that create relationship intimacy and perceived reciprocity among their followers (Lou 2021). Researchers have extended the notion of parasocial relationships to avatars, robots (Whang and Im 2021) and brands in general (Labrecque 2014). The frequency of engagement, including even low-level interactions, such as seeing others, helps in relationship development (Heide and Miner 1992). This is relevant in both real-life and online settings. Marketers draw on trust in parasocial relationships, for instance, in celebrities endorsing products (Chung and Cho 2017), which is well-received by consumers and increases their purchase intention to endorse products (Lou 2021). Nonetheless, developing a parasocial relationship also depends on perceived authenticity and credibility (Whang and Im 2021), including an organic fit between the influencer and the endorsed products/services. Perceived similarities strengthen relationships: humans are naturally drawn to others who have similar traits and characteristics, even if these similarities only exist at the level of perception and even if the subject of attraction is non-human (Whang and Im 2021). However, a non-human entity is typically perceived as genuine when it is anthropomorphized and human-like (Whang and Im 2021). Perceived mutuality is stronger with human-like non-humans (Whang and Im 2021), and consumers associate higher levels of consciousness and responsibility with them, as opposed to less human-like non-humans (Aggarwal and McGill 2007). Both the study of parasocial relationships and social comparison theory are concerned with relationships, social interactions, and self-worth. Parasocial relationships influence self-evaluation and perceived self-worth. For instance, when people compare themselves unfavorably to celebrity figures, their self-esteem may decrease (Staniewski and Awruk 2022). The existing literature in this area lacks an in-depth understanding of how anthropomorphization influences parasocial relationships and social comparisons between virtual influencers and their followers. The following section delves into the relevant aspects of social comparison theory. Social comparison theory Social comparison processes can influence identification; thus, followers may assess their position within a group of fans. This may be a human or non-human fan group; in either case, the parasocial relationships within the follower group can contribute to an individual’s social identity (Aw and Chuah 2021). Social comparison is a widely shared human trait that involves comparing ourselves to others and is an important element of human life. This theory explains how individuals compare themselves to others and gain or lose self-esteem, depending on evaluations of their abilities, achievements, beliefs, attitudes, and appearances compared to those of their peers (Festinger 1954). Advertising scholarship shows that social comparisons to idealized celebrities can lower one’s self-esteem related to physical attractiveness and financial success (Gulas and McKeage 2000), and increase the sharing of materialistic values enacted through imitation, such as the purchase of products that celebrities possess (Chan and Prendergast 2008). Goethals (1986) views social comparison theory as a natural extension of researching how people evaluate themselves, approach their reference groups, and the comparative functions reference groups provide, including pressure towards conformity. Suls and Wheeler (2013) identified influencing factors in social comparison theory, such as individuals’ adaptiveness, egocentrism, tendency for projecting, approaches to belief formation, and whether individuals critically revisit or review beliefs. Social comparisons may lead to both positive (e.g., feeling motivated to improve) and negative emotions. For instance, envy towards those perceived as more attractive may lead to undervaluing them in other areas, such as professional skills (Bower 2001).Incorporating social comparison theory into our theorization opens up versatile possibilities. Comparisons are inherent in relationships with others and in one’s own relationships with oneself. Researchers have used it to study burnout (Buunk and Schaufeli 2018) and deception regarding consumption (Argo et al. 2006). This theory identifies multiple ways of making comparisons. Comparisons can range from extremely negative (upward comparisons) to highly positive valence (downward comparisons) (Goethals 1986). Examples of negativity relate to a lack of selfappreciation, depressive feelings, distress, and offending others, whereas positivity is typically about feeling more motivated and grateful as a result of social comparisons. Despite their practical relevance, little is known about social comparisons that occur between humans in relation to virtual influencers. Hence, this study addresses the following research question: How does the virtual influencer phenomenon influence followers’ online social experiences through social comparisons and parasocial relationships? This study sought to uncover how social comparisons and parasocial relationships manifest in the virtual influencer phenomenon. It aims to further understand how individuals perceive themselves in comparison to virtual influencers and how they form emotional relationships with them. Utilizing a theory of knitting approach following De Pelecijn et al. (2023), synergies have been identified between the key theoretical streams of anthropomorphization, parasocial relationships, and social comparison theory:relevance, little is known about social comparisons that occur between humans in relation to virtual influencers. Hence, this study addresses the following research question: How does the virtual influencer phenomenon influence followers’ online social experiences through social comparisons and parasocial relationships? This study sought to uncover how social comparisons and parasocial relationships manifest in the virtual influencer phenomenon. It aims to further understand how individuals perceive themselves in comparison to virtual influencers and how they form emotional relationships with them. Utilizing a theory of knitting approach following De Pelecijn et al. (2023), synergies have been identified between the key theoretical streams of anthropomorphization, parasocial relationships, and social comparison theory:
考点1:”human influencer“推荐翻译为“真人网红”,因这篇文章的人工智能语境,人类与AI形成对比。翻译为网红、人类影响者均不对。 考点2:“virtual influencers”推荐翻译为“虚拟网红”,注意全文的人工智能营销、网红营销的语境。 考点3:”This development offers a fresh perspective on the role of influencers,”推荐翻译为“这一发展为研究网红的角色提供了新鲜视角,“ 考点4:”ecosystems.”社交媒体生态” 考点5:“virtual agents”应结合上下文语境和学界惯例翻译为“虚拟人” 考点6:“direction”推荐翻译为趋势 考点7:“enabling”推荐翻译为使XX成为可能 考点8:“ perceived anthropomorphism"推荐翻译为“感知拟人性(感知拟人度也可)“ 考点9:”attribute to ”推荐翻译为”将XXX赋予“。 考点10:”settings“推荐翻译为“情境” 考点11:”reveal“推荐翻译为“分享” 考点12:”avatars“推荐翻译为“虚拟形象” 考点13:”relevant“推荐翻译为“适用的” 考点14:”human or non-human fan group“应翻译为”崇拜真人或虚拟角色的粉丝群体“ 考点15: “increase the sharing of materialistic values enacted through imitation,“ 推荐翻译为”增加通过模仿来实现的物质主义价值观的传播“ 考点16: ”practical relevance“推荐翻译为”实际相关性“ 考点17:”key theoretical streams“推荐翻译为”关键理论脉络“ 考点18:“Nonhuman agents”推荐翻译为“非人类主体”
42d59
学术论文
社会科学
100
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Why learn English Learning English is a good way to improve your professional future and expand your horizons. More and more people around the world use English as a way to connect with people from different cultural backgrounds. The English language has become the lingua franca of international communication, trade, business, diplomacy, and many other areas. Mastering the language will open doors for you, both expected and unexpected. According to the EF English Proficiency Index (EPI), the world's largest ranking of countries and regions by English skills, more than 1 billion people speak English as a first or second language, and hundreds of millions as a third or fourth. English proficiency helps scientists, researchers, tourists, and business professionals exchange information. On an individual level, people who speak English get better jobs, earn more, and have access to more of the information available online. 6 benefits of learning English If you're still not convinced of the benefits of learning English, here are a few more: 1. Learning English will build up your confidence. Mastering something you've struggled with for years is extremely satisfying. 2. Learning English boosts your career by allowing you to follow the latest trends in your field, attend conferences, and evolve into positions with an international remit. 3. Learning a new language keeps your mind healthy, stimulating your memory and increasing mental flexibility. 4. Learning English gives you access to more of the world's knowledge, and from a wider variety of perspectives. Over half of the material online is in English, and all of the world's top scientific journals are published in that language. 5. Learning English makes international travel easier and more fulfilling. You'll be less dependent on guides and translators, able to explore the world on your own. 6. Learning English allows you to meet people from other places and expand your cultural understanding. The world is more interconnected than ever before. How to learn English The secret to learning English quickly isn’t much of a secret: increase your exposure to the language. Ideally, you should be surrounded by English 24 hours a day, 7 days a week. That’s why going abroad is such an effective way to learn a language. But if travel isn’t an option, you can still create an English immersion environment at home. Do it by reading in English, listening to English podcasts, watching TV in English, speaking to people in English at work, making English-speaking friends online, etc. At first, this sounds daunting. But you can do it if you take it one step at a time. 1. First, think of something you’re already doing every day. Let’s pretend you like to jog while listening to a sports podcast. 2. Now look for a sports podcast in English. Google is your friend. 3. When you jog tomorrow, listen to it. Congratulations, you’ve just converted 30 minutes of your day into English! Repeat this procedure to convert chunks of your day bit by bit. If you read the newspaper, switch to an English-language paper. If you watch football, watch the match on a British TV station. If you cook, use recipes from American websites. If you make pottery, find an English-language pottery forum where you discuss your craft. Everything you’re already doing can be done in English, guaranteed. You can also start to do new things in English too, of course. Do you have English-speaking colleagues? Send them friendly e-mails. Chat with them in the empty minutes before a meeting begins. Are there foreign tourists in your city? Find out if you can volunteer at one of the sights they visit: a museum, a park, a monument. Or maybe your airport or train station needs volunteers. You can also go hang out at one of those places informally and help people who look lost. Just be careful not to be creepy about it. Is there an expat community in your city? Look on Facebook for American, British, or Canadian cultural associations. They may have monthly or bi-monthly meetups you can attend. These tend to be informal events in a bar or restaurant and are open to everyone. How to stay motivated while learning English One of the most common mistakes people make when trying to learn English is underestimating how long it will take. Everyone says children are natural language learners, but they’re two years old before they even start to form sentences. Learning English is a long-term goal. Be patient with yourself. Staying motivated over months or years is a major challenge, but it’s essential if you want to master English. A huge body of research proves it. Here are a few tips to keep your motivation high: 1. Do things in English that you already like doing. If you’re having fun while speaking English, you’ll be less inclined to quit. Follow your passions. 2. Make English a habit. Build it in to your everyday life. The less you have to think about it, the more likely you are to use English every day. Automatically. 3. Keep a learning journal. Write down what you're doing to improve your English, and keep track of milestones like your first movie in English without subtitles, your first meeting in English, or your first date with an English speaker. 4. Award yourself prizes for accomplishing specific goals. The prizes can be big or small. A new shirt. A candy bar. A trip to London. Your brain is wired to detect which types of activities deliver rewards. If you make English one of those activities, it will automatically motivate you to learn English. 5. Keep track of the numbers. Our free standardized English test will help you. Take it every 6 months. How to measure your English level It is important to measure your English level before you start studying and then regularly throughout the learning process. Otherwise, you can't tell how much progress you're making. There are many different ways of measuring progress. The best way of establishing a benchmark is to test yourself against the goal you are trying to reach. For example, if your goal is to be able to make presentations at work in English, then you need to try and make a presentation at work in English and see how it goes. Take some notes afterwards about how you felt and what you need to work on. This will be your benchmark. After 6 months of learning English and working towards this goal, make another presentation and see how much you've improved. If you've got a more generic goal, taking a standardized English test(the TOEFL or IELTS) is a good way to get a benchmark of your current level and keep track of your progress moving forward.
Why learn English Learning English is a good way to improve your professional future and expand your horizons. More and more people around the world use English as a way to connect with people from different cultural backgrounds. The English language has become the lingua franca of international communication, trade, business, diplomacy, and many other areas. Mastering the language will open doors for you, both expected and unexpected. According to the EF English Proficiency Index (EPI), the world's largest ranking of countries and regions by English skills, more than 1 billion people speak English as a first or second language, and hundreds of millions as a third or fourth. English proficiency helps scientists, researchers, tourists, and business professionals exchange information. On an individual level, people who speak English get better jobs, earn more, and have access to more of the information available online. 6 benefits of learning English If you're still not convinced of the benefits of learning English, here are a few more: 1. Learning English will build up your confidence. Mastering something you've struggled with for years is extremely satisfying. 2. Learning English boosts your career by allowing you to follow the latest trends in your field, attend conferences, and evolve into positions with an international remit. 3. Learning a new language keeps your mind healthy, stimulating your memory and increasing mental flexibility. 4. Learning English gives you access to more of the world's knowledge, and from a wider variety of perspectives. Over half of the material online is in English, and all of the world's top scientific journals are published in that language. 5. Learning English makes international travel easier and more fulfilling. You'll be less dependent on guides and translators, able to explore the world on your own. 6. Learning English allows you to meet people from other places and expand your cultural understanding. The world is more interconnected than ever before. How to learn English The secret to learning English quickly isn’t much of a secret: increase your exposure to the language. Ideally, you should be surrounded by English 24 hours a day, 7 days a week. That’s why going abroad is such an effective way to learn a language. But if travel isn’t an option, you can still create an English immersion environment at home. Do it by reading in English, listening to English podcasts, watching TV in English, speaking to people in English at work, making English-speaking friends online, etc. At first, this sounds daunting. But you can do it if you take it one step at a time. 1. First, think of something you’re already doing every day. Let’s pretend you like to jog while listening to a sports podcast. 2. Now look for a sports podcast in English. Google is your friend. 3. When you jog tomorrow, listen to it. Congratulations, you’ve just converted 30 minutes of your day into English! Repeat this procedure to convert chunks of your day bit by bit. If you read the newspaper, switch to an English-language paper. If you watch football, watch the match on a British TV station. If you cook, use recipes from American websites. If you make pottery, find an English-language pottery forum where you discuss your craft. Everything you’re already doing can be done in English, guaranteed. You can also start to do new things in English too, of course. Do you have English-speaking colleagues? Send them friendly e-mails. Chat with them in the empty minutes before a meeting begins. Are there foreign tourists in your city? Find out if you can volunteer at one of the sights they visit: a museum, a park, a monument. Or maybe your airport or train station needs volunteers. You can also go hang out at one of those places informally and help people who look lost. Just be careful not to be creepy about it. Is there an expat community in your city? Look on Facebook for American, British, or Canadian cultural associations. They may have monthly or bi-monthly meetups you can attend. These tend to be informal events in a bar or restaurant and are open to everyone. How to stay motivated while learning English One of the most common mistakes people make when trying to learn English is underestimating how long it will take. Everyone says children are natural language learners, but they’re two years old before they even start to form sentences. Learning English is a long-term goal. Be patient with yourself. Staying motivated over months or years is a major challenge, but it’s essential if you want to master English. A huge body of research proves it. Here are a few tips to keep your motivation high: 1. Do things in English that you already like doing. If you’re having fun while speaking English, you’ll be less inclined to quit. Follow your passions. 2. Make English a habit. Build it in to your everyday life. The less you have to think about it, the more likely you are to use English every day. Automatically. 3. Keep a learning journal. Write down what you're doing to improve your English, and keep track of milestones like your first movie in English without subtitles, your first meeting in English, or your first date with an English speaker. 4. Award yourself prizes for accomplishing specific goals. The prizes can be big or small. A new shirt. A candy bar. A trip to London. Your brain is wired to detect which types of activities deliver rewards. If you make English one of those activities, it will automatically motivate you to learn English. 5. Keep track of the numbers. Our free standardized English test will help you. Take it every 6 months. How to measure your English level It is important to measure your English level before you start studying and then regularly throughout the learning process. Otherwise, you can't tell how much progress you're making. There are many different ways of measuring progress. The best way of establishing a benchmark is to test yourself against the goal you are trying to reach. For example, if your goal is to be able to make presentations at work in English, then you need to try and make a presentation at work in English and see how it goes. Take some notes afterwards about how you felt and what you need to work on. This will be your benchmark. After 6 months of learning English and working towards this goal, make another presentation and see how much you've improved. If you've got a more generic goal, taking a standardized English test(the TOEFL or IELTS) is a good way to get a benchmark of your current level and keep track of your progress moving forward.
考点1:“lingua franca”推荐译为“通用语”或“国际通用语言”。 考点2:“remit”在“positions with an international remit”中推荐译为“职责范围”或“工作范畴”。 考点3:“standardized English test”应保持“标准化英语测试”表达。 考点4:“EF English Proficiency Index (EPI)”应译为“英孚英语熟练度指标 (EPI)”
431b6
垂类场景
教育
22
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 从1968年的《太阳王子霍尔斯的大冒险》到2013年的《起风了》,不难发现,在其不同时期的动画作品中都体现着他独特的世界观、历史观、人生观和艺术观。并且,还构筑了一个令人叹为观止宫式世界,干净的画面配上时而空灵、时而轻快、时而震撼的音乐,让人不由自主的深陷其中,发现、体会、探索并深思其中暗藏的人生哲理。更值得惊叹的是,他的动画作品大多以坚强且独立的女孩或年轻女性为主角,以她们独特的视角反映不同题材的故事,把令人反思的讯息融入其中,用其独特的创作手法将环保、人生、梦想、生存等要素淋漓尽致的展现在作品中,成功的塑造了一个个活灵活现的人物,触动着我们的心灵。    一、宫崎骏影视动画作品的主题特色    (一)人文主义的深刻体现    人文主义是指社会价值取向倾向于对人的个性的关怀,注重强调维护人性尊严,提倡宽容,反对暴力,主张自由平等和自我价值体现的一种哲学思潮与世界观。宫崎骏的影视动画作品以关注环境、战争、人生、梦想、真情等为主题,将人与人、人与自然之间的关系作为主线展开叙述。其前期的作品主要表现人与自然之间的关系为主,后期的作品主要探讨人与人、社会之间的关系为主。《风之谷》、《龙猫》、《千与千寻》等都用人物细致入微的演绎展现她们心中极致的爱。    (二)女性主义的完美诠释    宫崎骏在其影视动画作品中善用女性的视角隐晦的去揭示那些被我们忽视的问题——人与自然的不和谐、环境的污染、战争的伤害等。很显然,这些主题恰恰是需要全人类甚至是全世界共同关注的。其影片中的女主角既有女性的柔美与温和,又有在面对困难与危险时的坚毅和刚强,很显然这些人物的塑造成为影片中美好与胜利的化身。《幽灵公主》、《哈尔的移动城堡》等作品中的女性形象用她们的行动诠释着生活的意义。    宫崎骏通过动画使人文主义和女性主义的主题得到最完美的诠释。在“宫崎骏世界”里,少女是纯洁美好的象征,透过一双真挚善良的双眸带领着观影者跟随导演一起感受影片中女性的成长历程,去体会生活中的真善美。影片将贴合故事情节的人文主义通过少女们感性细腻的瞳孔淋漓尽致的展现在观影者的面前。 二、宫崎骏影视动画作品的风格特色 (一)人物选择的象征性——用小人物塑造大形象   宫崎骏的影视动画作品中的人物大多是平凡的年轻女性,而且也打破了主角特有的模式,大多采用的是平凡无奇的面孔充当故事的主角。并且,通过故事情节的曲折发展和主角身上独有的魅力来吸引观众的注意。正因为有了这样的人物塑造与定位,很容易使观影者深入其中,并发现主角的魅力所在,也可以在最大程度上破除了影视作品中的主角与现实中人的距离。 (二)主题的多样性——用小事件展现大内涵   宫崎骏的影视动画作品中的事件几乎是身边随处可见。然而,宫崎骏为现实蒙上了写意的面纱,用细腻的笔法隐晦的描绘细节并展现深刻的内涵。影片中栩栩如生的人物,震撼着我们的心灵,使我们不得不反思自己在现实社会中所扮演的角色,更让我们深刻的反思自己的行为是否正确。   宫崎骏以其独特的宫式风格用小人物塑造大形象,用小事件展现大内涵,用最真实的情感述说故事,用最细腻的笔法描绘情节,起到了小中见大的作用。而恰恰是这些生活中的小事能够带给观影者心灵上的洗礼,并深受触动甚至产生共鸣。
从1968年的《太阳王子霍尔斯的大冒险》到2013年的《起风了》,不难发现,在其不同时期的动画作品中都体现着他独特的世界观、历史观、人生观和艺术观。并且,还构筑了一个令人叹为观止宫式世界,干净的画面配上时而空灵、时而轻快、时而震撼的音乐,让人不由自主的深陷其中,发现、体会、探索并深思其中暗藏的人生哲理。更值得惊叹的是,他的动画作品大多以坚强且独立的女孩或年轻女性为主角,以她们独特的视角反映不同题材的故事,把令人反思的讯息融入其中,用其独特的创作手法将环保、人生、梦想、生存等要素淋漓尽致的展现在作品中,成功的塑造了一个个活灵活现的人物,触动着我们的心灵。    一、宫崎骏影视动画作品的主题特色    (一)人文主义的深刻体现    人文主义是指社会价值取向倾向于对人的个性的关怀,注重强调维护人性尊严,提倡宽容,反对暴力,主张自由平等和自我价值体现的一种哲学思潮与世界观。宫崎骏的影视动画作品以关注环境、战争、人生、梦想、真情等为主题,将人与人、人与自然之间的关系作为主线展开叙述。其前期的作品主要表现人与自然之间的关系为主,后期的作品主要探讨人与人、社会之间的关系为主。《风之谷》、《龙猫》、《千与千寻》等都用人物细致入微的演绎展现她们心中极致的爱。    (二)女性主义的完美诠释    宫崎骏在其影视动画作品中善用女性的视角隐晦的去揭示那些被我们忽视的问题——人与自然的不和谐、环境的污染、战争的伤害等。很显然,这些主题恰恰是需要全人类甚至是全世界共同关注的。其影片中的女主角既有女性的柔美与温和,又有在面对困难与危险时的坚毅和刚强,很显然这些人物的塑造成为影片中美好与胜利的化身。《幽灵公主》、《哈尔的移动城堡》等作品中的女性形象用她们的行动诠释着生活的意义。    宫崎骏通过动画使人文主义和女性主义的主题得到最完美的诠释。在“宫崎骏世界”里,少女是纯洁美好的象征,透过一双真挚善良的双眸带领着观影者跟随导演一起感受影片中女性的成长历程,去体会生活中的真善美。影片将贴合故事情节的人文主义通过少女们感性细腻的瞳孔淋漓尽致的展现在观影者的面前。 二、宫崎骏影视动画作品的风格特色 (一)人物选择的象征性——用小人物塑造大形象   宫崎骏的影视动画作品中的人物大多是平凡的年轻女性,而且也打破了主角特有的模式,大多采用的是平凡无奇的面孔充当故事的主角。并且,通过故事情节的曲折发展和主角身上独有的魅力来吸引观众的注意。正因为有了这样的人物塑造与定位,很容易使观影者深入其中,并发现主角的魅力所在,也可以在最大程度上破除了影视作品中的主角与现实中人的距离。 (二)主题的多样性——用小事件展现大内涵   宫崎骏的影视动画作品中的事件几乎是身边随处可见。然而,宫崎骏为现实蒙上了写意的面纱,用细腻的笔法隐晦的描绘细节并展现深刻的内涵。影片中栩栩如生的人物,震撼着我们的心灵,使我们不得不反思自己在现实社会中所扮演的角色,更让我们深刻的反思自己的行为是否正确。   宫崎骏以其独特的宫式风格用小人物塑造大形象,用小事件展现大内涵,用最真实的情感述说故事,用最细腻的笔法描绘情节,起到了小中见大的作用。而恰恰是这些生活中的小事能够带给观影者心灵上的洗礼,并深受触动甚至产生共鸣。
考点1:“宫崎骏世界”和“宫式世界”是一个意思,应该采用统一的翻译 考点2:“生活中的真善美”推荐译为“the kindness and the beauty in daily life” 考点3:“故事情节的曲折发展“推荐译为“plenty of ups-and-downs” 考点4:“细腻的笔法”推荐译为“delicately crafted stories” 考点5:“人与人、社会之间的关系”推荐译为“the relationships between people and between individuals and society”,不可直译为“the relationships between people and society” 考点6:“人文主义”译为“humanism” 考点7:“女性主义”译为“feminism”
44011
学术论文
人文科学
18
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 迪士尼历经了“内容起家—渠道布局(线下乐园+线上广播)—回归内容—渠道再布局(流媒体)”的发展过程,形成了内容和渠道的深度绑定和协同。迪士尼自 1923 年成立之始专注于电影内容制作,打造了米老鼠、唐老鸭等知名 IP;1955 年开始陆续推出多个迪士尼现代主题公园,形成了面向全球的线下变现渠道;1981-2001 之间,迪士尼公布有线电视网络计划、成立电影发行商试金石影业、收购美国广播公司 ABC,完成线上渠道布局;2005 年起迪士尼收购皮克斯、漫威、卢卡斯等内容制作工作室,回归内容建设;2012 年迪士尼流媒体布局拉开大幕,开始向新的流媒体渠道延伸。2017 年迪士尼宣布将推出自有流媒体平台,并计划于 19 年将其全部电影从 Netflix下架而作为自有平台的独占内容;2018 年底,迪士尼自有流媒体平台命名为“Disney+”;19 年 4 月,迪士尼确认 Disney+于 19 年 11 月上线。加上 18 年已上线的体育流媒体平台ESPN+,迪士尼流媒体战略下的三把利剑基本打造完毕。 迪士尼加速进军流媒体受市场青睐。自 3 月完成对福克斯的收购后,迪士尼股价快速上行,截至 2019 年 6 月 5 日累计涨幅超过 20%,而同期奈飞股价则震荡下行,迪士尼加速进军流媒体为什么被市场看好?它的核心竞争优势在哪?迪士尼竞争核心:时间构筑内容壁垒、多变现渠道降低内容成本率、充裕现金流支撑内容投入。 1)迪士尼内容自制始于 1923 年,近百年的自制+收购历史绘制庞大内容版图。1)自制:迪士尼自 1923 年创立以来,塑造了米老鼠、唐老鸭等经典动画形象;1950 年代打造了《金银岛》、《加勒比海盗》等真人系列电影; 2)收购:2006 年以 74 亿美元并购皮克斯动画工作室,2009 年以 40 亿美元并购漫威(旗下有多达 5000 个漫画人物),2012 年以 40.5亿美元并购手握星球大战系列 IP 的卢卡斯影业,累计并购支出达到 156.9 亿美元,2019年 3 月以 713 亿美元收购 21 世纪福克斯。目前已形成华特迪士尼影业、漫威影业、皮克斯影业、卢卡斯影业、21 世纪福克斯等在内的完整内容体系。相比之下,自制历史较短、原创占比较低的 Netflix 内容投入大,且外购版权面临一定的可持续问题。1)从自制内容投入占比看,渠道起家的 Netflix 2012 年才开启原创内容制作,2015 年起其自制内容投入开始提速,至 1Q19 Netflix 原创内容资产占全部内容资产比例达到 31.5%。但相比之下,迪士尼长期保持较高的自制内容投入比例,其自制内容投入占内容总投入比例超过 40%(净现金流出口径),且即将上线的 Disney+平台上的内容将全部为迪士尼自有内容及相关衍生节目。2)从 Netflix 外购版权价格和持续性看:受到竞争对手排他策略、以及授权合同到期等因素影响,外购成本快速上升,从 2013 年4000 万美元一季的《铁杉树丛》到 2016 年 1.3 亿美元一季的《王冠》,头部内容的采购成本不断上涨。迪士尼 2017 年 8 月宣布从 2019 年起撤出奈飞平台上所有迪士尼电影版权,且 19 年开始不再向奈飞提供新的迪士尼和皮克斯电影,外购版权面临一定的可持续问题。 2)多变现渠道提高内容变现效率,较高利润率为迪士尼流媒体业务构建定价优势。传媒产品一次性生产成本高,降低成本有两种方式,一是尽可能压缩成本,提高效率;二是生产更多的商品,将固定成本摊销到更多的产品中去,比如对同一产品进行多次销售,复制过程中边际成本递减。迪士尼电影产品制作完成后,线上可通过广播、电视、流媒体平台变现,线下通过影院票房、主题乐园变现,此外版权分销及 IP 授权均可带来内容的多次变现;而 Netflix 目前仅有会员收入这一变现渠道。较高的内容变现效率带来较高利润率,客观上为迪士尼流媒体业务实现较低定价作支撑。较高利润率客观上支撑了迪士尼流媒体平台会员较低的定价。通过低价获取用户,扩大用户规模和会员收入,再反哺到内容制作中,形成正向循环。 3)迪士尼现金流充裕,且融资成本较低。迪士尼拥有足够的现金流支撑其流媒体布局(预计 2020 年为 Disney+投入 10 亿美元进行原创内容制作,年投入金额 2024 年将达到 25 亿美元),而 Netflix 则需要高额举债来投入原创。为维持用户增长,Netflix 近年在原创和外购版权内容上的投入迅速增加,2018 年 Netflix 内容成本净现金流出达 120.4 亿美元,占营收的 77.9%,2019 年 4 月计划发行 20 亿美元债券将使 Netflix 的长期债务达到约 123 亿美元,给公司造成较大现金流压力。
迪士尼历经了“内容起家—渠道布局(线下乐园+线上广播)—回归内容—渠道再布局(流媒体)”的发展过程,形成了内容和渠道的深度绑定和协同。迪士尼自 1923 年成立之始专注于电影内容制作,打造了米老鼠、唐老鸭等知名 IP;1955 年开始陆续推出多个迪士尼现代主题公园,形成了面向全球的线下变现渠道;1981-2001 之间,迪士尼公布有线电视网络计划、成立电影发行商试金石影业、收购美国广播公司 ABC,完成线上渠道布局;2005 年起迪士尼收购皮克斯、漫威、卢卡斯等内容制作工作室,回归内容建设;2012 年迪士尼流媒体布局拉开大幕,开始向新的流媒体渠道延伸。2017 年迪士尼宣布将推出自有流媒体平台,并计划于 19 年将其全部电影从 Netflix下架而作为自有平台的独占内容;2018 年底,迪士尼自有流媒体平台命名为“Disney+”;19 年 4 月,迪士尼确认 Disney+于 19 年 11 月上线。加上 18 年已上线的体育流媒体平台ESPN+,迪士尼流媒体战略下的三把利剑基本打造完毕。 迪士尼加速进军流媒体受市场青睐。自 3 月完成对福克斯的收购后,迪士尼股价快速上行,截至 2019 年 6 月 5 日累计涨幅超过 20%,而同期奈飞股价则震荡下行,迪士尼加速进军流媒体为什么被市场看好?它的核心竞争优势在哪?迪士尼竞争核心:时间构筑内容壁垒、多变现渠道降低内容成本率、充裕现金流支撑内容投入。 1)迪士尼内容自制始于 1923 年,近百年的自制+收购历史绘制庞大内容版图。1)自制:迪士尼自 1923 年创立以来,塑造了米老鼠、唐老鸭等经典动画形象;1950 年代打造了《金银岛》、《加勒比海盗》等真人系列电影; 2)收购:2006 年以 74 亿美元并购皮克斯动画工作室,2009 年以 40 亿美元并购漫威(旗下有多达 5000 个漫画人物),2012 年以 40.5亿美元并购手握星球大战系列 IP 的卢卡斯影业,累计并购支出达到 156.9 亿美元,2019年 3 月以 713 亿美元收购 21 世纪福克斯。目前已形成华特迪士尼影业、漫威影业、皮克斯影业、卢卡斯影业、21 世纪福克斯等在内的完整内容体系。相比之下,自制历史较短、原创占比较低的 Netflix 内容投入大,且外购版权面临一定的可持续问题。1)从自制内容投入占比看,渠道起家的 Netflix 2012 年才开启原创内容制作,2015 年起其自制内容投入开始提速,至 1Q19 Netflix 原创内容资产占全部内容资产比例达到 31.5%。但相比之下,迪士尼长期保持较高的自制内容投入比例,其自制内容投入占内容总投入比例超过 40%(净现金流出口径),且即将上线的 Disney+平台上的内容将全部为迪士尼自有内容及相关衍生节目。2)从 Netflix 外购版权价格和持续性看:受到竞争对手排他策略、以及授权合同到期等因素影响,外购成本快速上升,从 2013 年4000 万美元一季的《铁杉树丛》到 2016 年 1.3 亿美元一季的《王冠》,头部内容的采购成本不断上涨。迪士尼 2017 年 8 月宣布从 2019 年起撤出奈飞平台上所有迪士尼电影版权,且 19 年开始不再向奈飞提供新的迪士尼和皮克斯电影,外购版权面临一定的可持续问题。 2)多变现渠道提高内容变现效率,较高利润率为迪士尼流媒体业务构建定价优势。传媒产品一次性生产成本高,降低成本有两种方式,一是尽可能压缩成本,提高效率;二是生产更多的商品,将固定成本摊销到更多的产品中去,比如对同一产品进行多次销售,复制过程中边际成本递减。迪士尼电影产品制作完成后,线上可通过广播、电视、流媒体平台变现,线下通过影院票房、主题乐园变现,此外版权分销及 IP 授权均可带来内容的多次变现;而 Netflix 目前仅有会员收入这一变现渠道。较高的内容变现效率带来较高利润率,客观上为迪士尼流媒体业务实现较低定价作支撑。较高利润率客观上支撑了迪士尼流媒体平台会员较低的定价。通过低价获取用户,扩大用户规模和会员收入,再反哺到内容制作中,形成正向循环。 3)迪士尼现金流充裕,且融资成本较低。迪士尼拥有足够的现金流支撑其流媒体布局(预计 2020 年为 Disney+投入 10 亿美元进行原创内容制作,年投入金额 2024 年将达到 25 亿美元),而 Netflix 则需要高额举债来投入原创。为维持用户增长,Netflix 近年在原创和外购版权内容上的投入迅速增加,2018 年 Netflix 内容成本净现金流出达 120.4 亿美元,占营收的 77.9%,2019 年 4 月计划发行 20 亿美元债券将使 Netflix 的长期债务达到约 123 亿美元,给公司造成较大现金流压力。
考点1: 三把利剑,推荐翻译为three pillars或者three strategic platforms 考点2: 内容壁垒,推荐翻译为IP barriers 考点3: 内容版图,推荐翻译为IP library/content empire 考点4: 边际成本递减,必须翻译为diminishing marginal costs, 固定术语
441f7
新闻资讯
观点评论
155
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Thermal draw of the fibre computer The fibre draw process started with the construction of a macroscopic preform (Fig. 1b) comprising a sacrificial, stress-bearing polymethylmethacrylate (PMMA) cladding encasing an inner hollow thermoplastic cyclic olefin copolymer elastomer (ECOC) cylinder. During the draw process, the PMMA viscous fluid carried most of the tension and offered superior dimensional control, which the more rheologically complex viscoelastic fluid (ECOC) would not. Without this, the sudden transfer of tension from the elastomer to the electrodes as the preform collapses on devices leads to an asymmetric and non-conformal fibre profile near the devices. In addition, drawing the elastomer on its own under high stress would lead to molecular chain alignment, reducing the elasticity of the fibre once quenched. During the draw process, the wires with the pre-connected devices were fed into a single large channel of the preform while the dimensions were reduced to fibre size (Fig. 1b). To maintain the structural elasticity of the conductors through the draw, it is important that the helix diameter be substantially smaller than the fibre channel diameter to minimize the frictional resistance between the wires and the polymer. After the draw process, the PMMA layer was peeled away, leaving an ECOC fibre with eight encapsulated microelectronic devices separated by nearly 15 cm, connected by helically structured wires internal to the fibre (Extended Data Figs. 2 and 3). Because the size of some devices was larger than the steady-state diameter of the fibre away from the devices, which was about 1.35 mm, the channel of the fibre fully collapsed at the devices such that the devices were surrounded by a highly conformal layer of ECOC. Tensile and flexural properties Compared with planar approaches with anisotropic bending stiffness, this solvent-free high-speed thermal draw approach enables conformal, hermetic encapsulation of devices within fibres composed of an elastomer with more than an order of magnitude lower modulus than conventional thermoplastics28,31, providing isotropic flexibility and elasticity35. A representative tensile measurement of a section of fibre between devices (Fig. 1c) showed that breakage of the larger-diameter strain-limiting wire occurred at 6.6% strain, whereas the failure of the first electrode necessary for fibre computer functionality occurred at 67.1% strain. Similarly, a section of the fibre with an embedded device withstood 37% strain before failure (Supplementary Fig. 2). We also found that the fibres maintained functionality for more than a thousand cycles of stretching, representative of high yarn strain in conventional wearable textiles36,37,38, with further improvements in fatigue life possible through application-specific tuning of the wire helix properties (Extended Data Fig. 4). In addition to elastic deformations, we show that the fibre computer is capable of withstanding millimetre-scale flexural deformations, which are crucial for fabric formation and knitting in particular39,40. Bending radius measurements (Fig. 1d) showed that non-device-containing segments sustained flexural stresses down to a bending radius of 0.25 mm without failure, whereas a device-containing segment sustained flexural stresses down to a bending radius of 2.38 mm before the first break of a functional electrode due to the modulus mismatch near devices. We also found that the fibre computer withstood ten machine-washing cycles with no degradation in functionality (Extended Data Fig. 5), demonstrating environmental resilience along with mechanical reliability. Braided fibre cladding To further improve the mechanical robustness and surface properties of the fibre computers for fabric integration, we covered the fibre computers with diagonally interlaced textile yarns using a 16-carrier braiding machine (Fig. 1b). The elasticity of the fibre was maintained during the braiding process using a tensioned core loading approach. The versatility of this method with respect to reconfiguration of the type of yarn used, number of carriers and helix pitch allowed the independent design of the mechanical and surface properties of the fibre41,42. A wide variety of yarns were used to cover the fibre computers, including polylactic acid (PLA), Kevlar, cotton and polyester (Fig. 1e,f). The tensile strength of the fibre computer varied with yarn type and denier. A fibre computer braided with 360-denier cotton yarn had a tensile strength of 49 N, whereas a fibre computer braided with 540-denier polyester yarn had a tensile strength of 146 N (Extended Data Fig. 6). Beyond improvements in tensile strength, covering the core with low-friction yarn materials, such as polyester, reduced stiction compared to bare ECOC fibres and prevented stiction to surfaces, such as knitting needles, from causing the fibre to stretch beyond its failure strain. Fibre computer characterization Each fibre computer consists of eight devices connected on a single I2C bus, enabling sensing, computation, storage and communication in a single fibre (Fig. 2a). Four of these devices are sensors: a photodetector, a temperature sensor, a PPG sensor, and an accelerometer, which collect information from the environment and wearer through optical, thermal and mechanical transduction (Fig. 2b). The 32-bit MCU, capable of operating at frequencies up to 96 MHz, provides both computation and storage, with an ARM Cortex-M4 processor with a floating-point unit alongside 256 KB of onboard flash memory and 96 KB of RAM. If required, the generalizability of the interposer approach also allows storage capacity to be expanded by adding downstream memory chips. We implemented two electromagnetic methods for external communications: optically through an LED and through radio-frequency signals from a BLE unit. Intrafibre communication within the system uses the I2C addressing protocol, operating at a clock speed of 100 kHz, although the MCU can drive the bus frequencies up to a maximum of 3.4 MHz for higher throughput applications. The physical architecture uses four wires allocated for I2C (data, clock, power and ground) along with an additional wire dedicated to modulating the LED for optical communications. Through the I2C interface, devices along the length of the fibre can be individually addressed by means of unique 7-bit serial addresses, enabling processor interfacing with less than a 1-ms delay between sequential reads or writes (Extended Data Fig. 5). The MCU can also be dynamically reprogrammed over I2C to target different application requirements, even after fibre integration. The I2C protocol was selected for its use of only two signal conductors, minimizing potential mechanical failure modes; other protocols, such as serial peripheral interface (SPI) and universal asynchronous receiver-transmitter (UART), may also be adopted by either adding additional conductors to the fibre or through device-to-device communications on a single interposer. For applications with more stringent timing requirements, we found that transmission through the in-fibre electrodes over device-to-device spacings can support frequencies of up to 128 MHz (Supplementary Fig. 4), allowing for the adoption of higher-speed protocols if needed. Physiological computing demonstration To demonstrate the sensing, processing and communication capabilities of the fibre computer in a simple application, we show that health metrics, such as heart rate measurements, can be analysed on the basis of data collected by the embedded PPG sensor in close contact with the subject’s skin using signal filtering and feature recognition techniques applied by the in-fibre processor. As PPG sensors are susceptible to motion artefact noise, we show that the in-fibre MCU is capable of applying low-pass filtering and gradient thresholding to recognize individual pulses in real time from the PPG data stream received over I2C (Fig. 2c(ii)). The timing of the pulse peaks was used to calculate the wearer’s heart rate, and the MCU relayed these real-time estimates over I2C to the in-fibre BLE node, which was able to communicate the results to any BLE-compatible display device (Fig. 2c(iii)). Power considerations To power the embedded devices in the fibre computer, we integrated a pin-type lithium-ion battery only 3.65 mm in diameter inside an ECOC fibre, using a lower-temperature fibre draw to prevent the temperature rise from causing battery degradation. To allow for ease of connection for power, reprogramming and charging, the fibre computers and fibre batteries were terminated with 2.5-mm tip–ring–ring–sleeve (TRRS) audio connectors that can be connected together (Supplementary Fig. 7). After the draw process, we found that the battery-embedded fibre retained a rated battery capacity of 15 mAh (Fig. 2d(i)) and remained rechargeable over multiple uses (Fig. 2d(ii)). To reduce the battery voltage to the required 3.3 V power and I2C logic levels, an in-fibre LDO voltage regulator and pull-up resistors for the I2C bus were incorporated into the fibre computers. Given the baseline average current consumption during code execution (Fig. 2d(iii)), the battery-embedded fibre was capable of powering the fibre computer for up to 6 h. The capability for high-efficiency 1C discharge also demonstrated the ability to accommodate spikes in the current draw, such as those that occur during communications (Supplementary Fig. 6). Fabric networking Using the fibre computer, we proceeded to create a network of fibre computers in textile fabric. We chose optical and radio-frequency communications to avoid rigid interconnects and complex wiring schemes, which do not support ad hoc networking. Optical communications Optical communications were enabled through the use of a light sensor and LED present in each fibre. On–off keying modulation of the LED was used with Manchester encoding to eliminate the need for a separate clock signal and provide noise tolerance (Fig. 3c). We demonstrate optical communications up to 10 kHz (Extended Data Fig. 7); higher-speed optical communications can also be implemented using higher-bandwidth photodetectors. For multi-fibre optical communications, an addressable message protocol was established. The network architecture used a two-bus topology (Fig. 3a), where LED and light sensor pairs in each fibre were coupled to those in other fibres by means of two waveguide buses, which acted to improve the bidirectional communication distance compared to free-space transmission. For fibre pairs with LEDs that are aligned on the same bus, or if the distance between fibres is too large for detection, the message must be transmitted to an intermediary fibre, which can then repeat the message to the receiving fibre. To access the guided modes of the square PMMA waveguides, light must pass directly through the ECOC–PMMA interface without escaping into air. This requires a sufficient interfacial contact area between the fibre and waveguide, which can be created by deforming the surface of the elastomeric fibre with compression of the waveguide onto the fibre (Supplementary Figs. 8 and 9). We indeed found that transmission from one fibre to another was aided by increasing the compression force (Fig. 3d), with the amount of compression necessary varying on the basis of the waveguide size (Supplementary Fig. 10). At constant compression, we found that the light propagating in the waveguide had an attenuation coefficient of 0.18–0.28 cm−1 (Supplementary Fig. 10), with losses caused by the bending of the waveguide and out-coupling to the yarns of the fabric (Fig. 3e). In the future, alternate strategies for waveguide coupling and confinement, including grating couplers43,44, low-index claddings and directional light emitters, could be explored for higher-efficiency transmission.
Thermal draw of the fibre computer The fibre draw process started with the construction of a macroscopic preform (Fig. 1b) comprising a sacrificial, stress-bearing polymethylmethacrylate (PMMA) cladding encasing an inner hollow thermoplastic cyclic olefin copolymer elastomer (ECOC) cylinder. During the draw process, the PMMA viscous fluid carried most of the tension and offered superior dimensional control, which the more rheologically complex viscoelastic fluid (ECOC) would not. Without this, the sudden transfer of tension from the elastomer to the electrodes as the preform collapses on devices leads to an asymmetric and non-conformal fibre profile near the devices. In addition, drawing the elastomer on its own under high stress would lead to molecular chain alignment, reducing the elasticity of the fibre once quenched. During the draw process, the wires with the pre-connected devices were fed into a single large channel of the preform while the dimensions were reduced to fibre size (Fig. 1b). To maintain the structural elasticity of the conductors through the draw, it is important that the helix diameter be substantially smaller than the fibre channel diameter to minimize the frictional resistance between the wires and the polymer. After the draw process, the PMMA layer was peeled away, leaving an ECOC fibre with eight encapsulated microelectronic devices separated by nearly 15 cm, connected by helically structured wires internal to the fibre (Extended Data Figs. 2 and 3). Because the size of some devices was larger than the steady-state diameter of the fibre away from the devices, which was about 1.35 mm, the channel of the fibre fully collapsed at the devices such that the devices were surrounded by a highly conformal layer of ECOC. Tensile and flexural properties Compared with planar approaches with anisotropic bending stiffness, this solvent-free high-speed thermal draw approach enables conformal, hermetic encapsulation of devices within fibres composed of an elastomer with more than an order of magnitude lower modulus than conventional thermoplastics28,31, providing isotropic flexibility and elasticity35. A representative tensile measurement of a section of fibre between devices (Fig. 1c) showed that breakage of the larger-diameter strain-limiting wire occurred at 6.6% strain, whereas the failure of the first electrode necessary for fibre computer functionality occurred at 67.1% strain. Similarly, a section of the fibre with an embedded device withstood 37% strain before failure (Supplementary Fig. 2). We also found that the fibres maintained functionality for more than a thousand cycles of stretching, representative of high yarn strain in conventional wearable textiles36,37,38, with further improvements in fatigue life possible through application-specific tuning of the wire helix properties (Extended Data Fig. 4). In addition to elastic deformations, we show that the fibre computer is capable of withstanding millimetre-scale flexural deformations, which are crucial for fabric formation and knitting in particular39,40. Bending radius measurements (Fig. 1d) showed that non-device-containing segments sustained flexural stresses down to a bending radius of 0.25 mm without failure, whereas a device-containing segment sustained flexural stresses down to a bending radius of 2.38 mm before the first break of a functional electrode due to the modulus mismatch near devices. We also found that the fibre computer withstood ten machine-washing cycles with no degradation in functionality (Extended Data Fig. 5), demonstrating environmental resilience along with mechanical reliability. Braided fibre cladding To further improve the mechanical robustness and surface properties of the fibre computers for fabric integration, we covered the fibre computers with diagonally interlaced textile yarns using a 16-carrier braiding machine (Fig. 1b). The elasticity of the fibre was maintained during the braiding process using a tensioned core loading approach. The versatility of this method with respect to reconfiguration of the type of yarn used, number of carriers and helix pitch allowed the independent design of the mechanical and surface properties of the fibre41,42. A wide variety of yarns were used to cover the fibre computers, including polylactic acid (PLA), Kevlar, cotton and polyester (Fig. 1e,f). The tensile strength of the fibre computer varied with yarn type and denier. A fibre computer braided with 360-denier cotton yarn had a tensile strength of 49 N, whereas a fibre computer braided with 540-denier polyester yarn had a tensile strength of 146 N (Extended Data Fig. 6). Beyond improvements in tensile strength, covering the core with low-friction yarn materials, such as polyester, reduced stiction compared to bare ECOC fibres and prevented stiction to surfaces, such as knitting needles, from causing the fibre to stretch beyond its failure strain. Fibre computer characterization Each fibre computer consists of eight devices connected on a single I2C bus, enabling sensing, computation, storage and communication in a single fibre (Fig. 2a). Four of these devices are sensors: a photodetector, a temperature sensor, a PPG sensor, and an accelerometer, which collect information from the environment and wearer through optical, thermal and mechanical transduction (Fig. 2b). The 32-bit MCU, capable of operating at frequencies up to 96 MHz, provides both computation and storage, with an ARM Cortex-M4 processor with a floating-point unit alongside 256 KB of onboard flash memory and 96 KB of RAM. If required, the generalizability of the interposer approach also allows storage capacity to be expanded by adding downstream memory chips. We implemented two electromagnetic methods for external communications: optically through an LED and through radio-frequency signals from a BLE unit. Intrafibre communication within the system uses the I2C addressing protocol, operating at a clock speed of 100 kHz, although the MCU can drive the bus frequencies up to a maximum of 3.4 MHz for higher throughput applications. The physical architecture uses four wires allocated for I2C (data, clock, power and ground) along with an additional wire dedicated to modulating the LED for optical communications. Through the I2C interface, devices along the length of the fibre can be individually addressed by means of unique 7-bit serial addresses, enabling processor interfacing with less than a 1-ms delay between sequential reads or writes (Extended Data Fig. 5). The MCU can also be dynamically reprogrammed over I2C to target different application requirements, even after fibre integration. The I2C protocol was selected for its use of only two signal conductors, minimizing potential mechanical failure modes; other protocols, such as serial peripheral interface (SPI) and universal asynchronous receiver-transmitter (UART), may also be adopted by either adding additional conductors to the fibre or through device-to-device communications on a single interposer. For applications with more stringent timing requirements, we found that transmission through the in-fibre electrodes over device-to-device spacings can support frequencies of up to 128 MHz (Supplementary Fig. 4), allowing for the adoption of higher-speed protocols if needed. Physiological computing demonstration To demonstrate the sensing, processing and communication capabilities of the fibre computer in a simple application, we show that health metrics, such as heart rate measurements, can be analysed on the basis of data collected by the embedded PPG sensor in close contact with the subject’s skin using signal filtering and feature recognition techniques applied by the in-fibre processor. As PPG sensors are susceptible to motion artefact noise, we show that the in-fibre MCU is capable of applying low-pass filtering and gradient thresholding to recognize individual pulses in real time from the PPG data stream received over I2C (Fig. 2c(ii)). The timing of the pulse peaks was used to calculate the wearer’s heart rate, and the MCU relayed these real-time estimates over I2C to the in-fibre BLE node, which was able to communicate the results to any BLE-compatible display device (Fig. 2c(iii)). Power considerations To power the embedded devices in the fibre computer, we integrated a pin-type lithium-ion battery only 3.65 mm in diameter inside an ECOC fibre, using a lower-temperature fibre draw to prevent the temperature rise from causing battery degradation. To allow for ease of connection for power, reprogramming and charging, the fibre computers and fibre batteries were terminated with 2.5-mm tip–ring–ring–sleeve (TRRS) audio connectors that can be connected together (Supplementary Fig. 7). After the draw process, we found that the battery-embedded fibre retained a rated battery capacity of 15 mAh (Fig. 2d(i)) and remained rechargeable over multiple uses (Fig. 2d(ii)). To reduce the battery voltage to the required 3.3 V power and I2C logic levels, an in-fibre LDO voltage regulator and pull-up resistors for the I2C bus were incorporated into the fibre computers. Given the baseline average current consumption during code execution (Fig. 2d(iii)), the battery-embedded fibre was capable of powering the fibre computer for up to 6 h. The capability for high-efficiency 1C discharge also demonstrated the ability to accommodate spikes in the current draw, such as those that occur during communications (Supplementary Fig. 6). Fabric networking Using the fibre computer, we proceeded to create a network of fibre computers in textile fabric. We chose optical and radio-frequency communications to avoid rigid interconnects and complex wiring schemes, which do not support ad hoc networking. Optical communications Optical communications were enabled through the use of a light sensor and LED present in each fibre. On–off keying modulation of the LED was used with Manchester encoding to eliminate the need for a separate clock signal and provide noise tolerance (Fig. 3c). We demonstrate optical communications up to 10 kHz (Extended Data Fig. 7); higher-speed optical communications can also be implemented using higher-bandwidth photodetectors. For multi-fibre optical communications, an addressable message protocol was established. The network architecture used a two-bus topology (Fig. 3a), where LED and light sensor pairs in each fibre were coupled to those in other fibres by means of two waveguide buses, which acted to improve the bidirectional communication distance compared to free-space transmission. For fibre pairs with LEDs that are aligned on the same bus, or if the distance between fibres is too large for detection, the message must be transmitted to an intermediary fibre, which can then repeat the message to the receiving fibre. To access the guided modes of the square PMMA waveguides, light must pass directly through the ECOC–PMMA interface without escaping into air. This requires a sufficient interfacial contact area between the fibre and waveguide, which can be created by deforming the surface of the elastomeric fibre with compression of the waveguide onto the fibre (Supplementary Figs. 8 and 9). We indeed found that transmission from one fibre to another was aided by increasing the compression force (Fig. 3d), with the amount of compression necessary varying on the basis of the waveguide size (Supplementary Fig. 10). At constant compression, we found that the light propagating in the waveguide had an attenuation coefficient of 0.18–0.28 cm−1 (Supplementary Fig. 10), with losses caused by the bending of the waveguide and out-coupling to the yarns of the fabric (Fig. 3e). In the future, alternate strategies for waveguide coupling and confinement, including grating couplers43,44, low-index claddings and directional light emitters, could be explored for higher-efficiency transmission.
考点1:“thermal draw”应该译为热拉制”,因为这是面料行业的主流翻译 考点2:“polyester”必须译为“聚酯纤维/涤纶”,译文取其一就可,且译文前后需保持一致 考点3: "denier"必须译为旦/旦尼尔;丹/丹尼,译文取其一就可,且要保持译文一致性 考点4:“Power considerations”推荐译为“功率考虑”或者“功耗考虑”,因为整个段落在讲的就是电力能源的消耗,译文取其一就可,且要保持译文一致性 考点5:“Ad hoc networking 应该译为“自组织网络”,不可遗漏Ad hoc 的翻译,表示自行组织,或者临时的。 考点6:”message protocol“结合上下文应该译为“通信协议”,不可译为“消息协议”这种模糊用词 考点7:”PPG sensor“应该译为“光电容积脉搏波传感器”,首次出现时,译文应该给出全称,之后再给出简称 考点8:”I2C“的中文标准译名为 "I2C总线",全称 "Inter-Integrated Circuit",应该译为”集成电路间总线 或 内部集成电路总线“,译文选取一种即可,维持译文的一致性,首次出现时,译文应该给出全称,之后再给出简称 考点9:“serial peripheral interface (SPI)“应该译为译为”串行外设接口(简称 SPI接口)”首次出现时,译文应该给出全称,之后再给出简称 考点10:“universal asynchronous receiver-transmitter“应该译为”通用异步收发器(简称 UART)”,业内主流翻译 考点11:“RAM” 应该译为”随机存取存储器”,首次出现时,译文应该给出全称,之后再给出简称 考点12:“fibre computer”推荐译为“纤维计算机”或者“智能纤维计算机”,选取一种即可,保持译文一致性
44afb
学术论文
自然科学
200
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Liability, Indemnification and Release 1. Each party shall indemnify the other from any and all losses that may arise out of breach by such party of any of the warranties set forth in this Article, and each party shall indemnify the other party from any and all losses that may arise out of breach by such party of any of the agreed terms in this Agreement. 2. Without limiting any right or remedy available to the non-defaulting party at law or in equity, upon the termination of this Agreement in accordance with Articles X, the defaulting party shall indemnify the other party for all damages, costs, charges and expenses suffered or incurred by it in connection with the termination due to the negligence, breach of duty or other default or wrongdoing of the defaulting party, its servants, employees,agents or contractors. 3. The CJV shall be solely and fully responsible for the quality of the Products manufactured hereunder, for their conformity with the Technical Data and for their compliance with the laws and regulations from time to time in force in the Territory. The CJV shall indemnify Party B against any loss or damage directly or indirectly suffered by Party B as a result of the failure of the Products manufactured hereunder to comply with the Technical Data("Defective Products")or to comply with such laws or regulations; provided, however, that such indemnification shall not exceed the total of ex-factory sales price, costs of delivery and transportation and other costs associated with the recall of these Defective Products. 4. Each Party hereby indemnifies the other Party and undertakes to hold harmless and defend the other Party against any and all claims, suits, losses,damages, disbursements (including legal and management costs) arising out of any alleged or actual breach or failure to comply with the terms and conditions hereof including but not limited to any infringement of the otherParty's intellectual property or other rights occurring as a result of theoffending Party's fault, omission or activities in connection with the Project.5. Consultant hereby agrees to indemnify, hold harmless and defend Client from and against any and all claims, liabilities, losses, expenses (including reasonable attorneys' fees),or damages (collectively "Liabilities") asserted against Client by a third party to the extent such Liabilities result from the infringement of the Works delivered on any third party's trade secret,trademark, service mark, copyright or patent issued as of the date of this Agreement (collectively, an "Intellectual Property Right"); provided that Client: (i) promptly notifies Consultant of any third party claim subject to indemnification hereunder, (ii) gives Consultant the right to control and direct the preparation of a defense, the defense and any settlement of any such claim, (iii) gives full cooperation to Consultant for the defense of same and (iv) complies with Consultant's direction to cease any use of the Works which, in Consultant's sole judgment, is likely to be ruled an infringement on a third party's Intellectual Property Right.. 6. Each Party forever releases and discharges the other from all claims.debts, allegations, actions, causes of action and demands, whether known or unknown, arising from or in connection with the Claim and existing as at the date of this Settlement Agreement, including without limitation any liability for legal costs connected with or arising out of the subject matter of the Claim. Any other claims unrelated to the Claim that one Party may have against another, whether prior to or after the date hereof, shall not be affected or otherwise prejudiced by this Agreement. Party B irrevocably waives all rights that it may have in law or contract against Party A, its affiliates, subsidiaries and related companies howsoever arising in respect of or in connection with the Claim. Taxation 1. All taxes, customs duties and other charges payable (i) in connection with the importation of the Equipment to China and (ii) in the execution of this Agreement to be levied by the Government of the PRC on Party A shall be paid by Party A. 2. "Taxation" means: a) any liability of the Company to any form of taxation whether created or imposed in the PRC or any other part of the world and, without prejudice to the generality of the foregoing, includes any tax computed on profits or income, any tax computed on capital assets, estate duty, profits tax,provisional profits tax, interest tax, death duty, gift duty, payroll tax withholding tax, rates, custom and excise duty, transfer tax, inheritance tax,stamp duty, capital duty, employment taxes, value added tax, and generally any tax, imposition, levy or rates or any amount payable to the revenue.customs or fiscal authorities in any part of the world; and b) all costs, interest, penalties, charges and expenses incidental or relating to the liability referred to in (a) above. 3. All fees, expenses and other charges payable to Consultant hereunder do not include any sales, use, excise, value added or other applicable taxes,tariffs or duties, payment of which shall be the sole responsibility of Client(excluding any applicable taxes based on Consultant's net income or taxes arising from the employment or independent contractor relationship between Consultant and its personnel). In the event that such taxes, tariffs or duties are assessed against Consultant, Client shall reimburse Consultant for any such amounts paid by Consultant or, prior to the payment of such amounts by Consultant, provide Consultant with valid tax exemption certificates with respect thereto. If Client is required by law to make any tax deduction,withholding or payment from any amount paid or payable by Client to Consultant under this Agreement, the amount paid or payable to Consultant shall be grossed-up to the extent necessary to ensure that Consultant receives and retains, free of liability, the net amount equal to the amount that Consultant would have received and retained,had no tax deduction or withholding been made. 4.It is understood and agreed specifically by the parties to this Agreement that Party B shall be an independent contractor, and that nothing in this Agreement or in the performance of any of its provisions is intended or shall be construed to constitute either party an agent, legal representative.subsidiary, joint venturer, partner, employer, or employee of the other for any purpose whatsoever. As an independent contractor, Party B agrees that it will be responsible for and will pay all income and other national, federal.state, provincial, and local taxes and charges assessable on it. It is understood and agreed that nothing in this Agreement authori-zes Party B to make any contract, agreement, warranty or representation on Party A's behalf, or to incur any debt or other obligation in Party A's name, and that Party A shall in no event assume liability or be deemed liable hereunder as a result of any such action of Party B. 1. "Taxes" include all forms of taxation,estate duties, deductions,withholdings, duties, imposts,levies, fees, charges, social security contributions and rates imposed, levied, collected, withheld or assessed by any local, municipal, regional, urban, governmental, state, federal or other body in Hong Kong and elsewhere, and any interest, additional taxation,penalty, surcharge or fine in connection therewith.2. Consultant is performing the Services as an independent contractor and not as an employee of Client and none of Consultant's personnel shall be entitled to receive any compensation, benefits or other incidents0remployment from Client. Subject to Article X,Consultant shall be responsible for all taxes and other expenses arising from the employment or independent contractor relationship between Consultant and its personnel and the rendition of Services here-under by such personnel to Client.Nothing in this Agreement shall be deemed to constitute a partnership,joint venture, or fiduciary relationship between Client and Consultant, nor shall anything in this Agreement be deemed to create an agency relationship between Consultant and Client. Neither Consultant nor Client shall be or become liable or bound by any representation, act or omission whatsoever of the other. If any portion, term or provision of this Agreement shall be held illegal void or ineffective under, or in conflict with, any applicable law, the validity of the remaining provisions shall not be affected thereby, and a substitute provision will be negotiated to preserve as near as possible the original.intent of the Agreement.
Liability, Indemnification and Release 1. Each party shall indemnify the other from any and all losses that may arise out of breach by such party of any of the warranties set forth in this Article, and each party shall indemnify the other party from any and all losses that may arise out of breach by such party of any of the agreed terms in this Agreement. 2. Without limiting any right or remedy available to the non-defaulting party at law or in equity, upon the termination of this Agreement in accordance with Articles X, the defaulting party shall indemnify the other party for all damages, costs, charges and expenses suffered or incurred by it in connection with the termination due to the negligence, breach of duty or other default or wrongdoing of the defaulting party, its servants, employees,agents or contractors. 3. The CJV shall be solely and fully responsible for the quality of the Products manufactured hereunder, for their conformity with the Technical Data and for their compliance with the laws and regulations from time to time in force in the Territory. The CJV shall indemnify Party B against any loss or damage directly or indirectly suffered by Party B as a result of the failure of the Products manufactured hereunder to comply with the Technical Data("Defective Products")or to comply with such laws or regulations; provided, however, that such indemnification shall not exceed the total of ex-factory sales price, costs of delivery and transportation and other costs associated with the recall of these Defective Products. 4. Each Party hereby indemnifies the other Party and undertakes to hold harmless and defend the other Party against any and all claims, suits, losses,damages, disbursements (including legal and management costs) arising out of any alleged or actual breach or failure to comply with the terms and conditions hereof including but not limited to any infringement of the otherParty's intellectual property or other rights occurring as a result of theoffending Party's fault, omission or activities in connection with the Project.5. Consultant hereby agrees to indemnify, hold harmless and defend Client from and against any and all claims, liabilities, losses, expenses (including reasonable attorneys' fees),or damages (collectively "Liabilities") asserted against Client by a third party to the extent such Liabilities result from the infringement of the Works delivered on any third party's trade secret,trademark, service mark, copyright or patent issued as of the date of this Agreement (collectively, an "Intellectual Property Right"); provided that Client: (i) promptly notifies Consultant of any third party claim subject to indemnification hereunder, (ii) gives Consultant the right to control and direct the preparation of a defense, the defense and any settlement of any such claim, (iii) gives full cooperation to Consultant for the defense of same and (iv) complies with Consultant's direction to cease any use of the Works which, in Consultant's sole judgment, is likely to be ruled an infringement on a third party's Intellectual Property Right.. 6. Each Party forever releases and discharges the other from all claims.debts, allegations, actions, causes of action and demands, whether known or unknown, arising from or in connection with the Claim and existing as at the date of this Settlement Agreement, including without limitation any liability for legal costs connected with or arising out of the subject matter of the Claim. Any other claims unrelated to the Claim that one Party may have against another, whether prior to or after the date hereof, shall not be affected or otherwise prejudiced by this Agreement. Party B irrevocably waives all rights that it may have in law or contract against Party A, its affiliates, subsidiaries and related companies howsoever arising in respect of or in connection with the Claim. Taxation 1. All taxes, customs duties and other charges payable (i) in connection with the importation of the Equipment to China and (ii) in the execution of this Agreement to be levied by the Government of the PRC on Party A shall be paid by Party A. 2. "Taxation" means: a) any liability of the Company to any form of taxation whether created or imposed in the PRC or any other part of the world and, without prejudice to the generality of the foregoing, includes any tax computed on profits or income, any tax computed on capital assets, estate duty, profits tax,provisional profits tax, interest tax, death duty, gift duty, payroll tax withholding tax, rates, custom and excise duty, transfer tax, inheritance tax,stamp duty, capital duty, employment taxes, value added tax, and generally any tax, imposition, levy or rates or any amount payable to the revenue.customs or fiscal authorities in any part of the world; and b) all costs, interest, penalties, charges and expenses incidental or relating to the liability referred to in (a) above. 3. All fees, expenses and other charges payable to Consultant hereunder do not include any sales, use, excise, value added or other applicable taxes,tariffs or duties, payment of which shall be the sole responsibility of Client(excluding any applicable taxes based on Consultant's net income or taxes arising from the employment or independent contractor relationship between Consultant and its personnel). In the event that such taxes, tariffs or duties are assessed against Consultant, Client shall reimburse Consultant for any such amounts paid by Consultant or, prior to the payment of such amounts by Consultant, provide Consultant with valid tax exemption certificates with respect thereto. If Client is required by law to make any tax deduction,withholding or payment from any amount paid or payable by Client to Consultant under this Agreement, the amount paid or payable to Consultant shall be grossed-up to the extent necessary to ensure that Consultant receives and retains, free of liability, the net amount equal to the amount that Consultant would have received and retained,had no tax deduction or withholding been made. 4.It is understood and agreed specifically by the parties to this Agreement that Party B shall be an independent contractor, and that nothing in this Agreement or in the performance of any of its provisions is intended or shall be construed to constitute either party an agent, legal representative.subsidiary, joint venturer, partner, employer, or employee of the other for any purpose whatsoever. As an independent contractor, Party B agrees that it will be responsible for and will pay all income and other national, federal.state, provincial, and local taxes and charges assessable on it. It is understood and agreed that nothing in this Agreement authori-zes Party B to make any contract, agreement, warranty or representation on Party A's behalf, or to incur any debt or other obligation in Party A's name, and that Party A shall in no event assume liability or be deemed liable hereunder as a result of any such action of Party B. 1. "Taxes" include all forms of taxation,estate duties, deductions,withholdings, duties, imposts,levies, fees, charges, social security contributions and rates imposed, levied, collected, withheld or assessed by any local, municipal, regional, urban, governmental, state, federal or other body in Hong Kong and elsewhere, and any interest, additional taxation,penalty, surcharge or fine in connection therewith.2. Consultant is performing the Services as an independent contractor and not as an employee of Client and none of Consultant's personnel shall be entitled to receive any compensation, benefits or other incidents0remployment from Client. Subject to Article X,Consultant shall be responsible for all taxes and other expenses arising from the employment or independent contractor relationship between Consultant and its personnel and the rendition of Services here-under by such personnel to Client.Nothing in this Agreement shall be deemed to constitute a partnership,joint venture, or fiduciary relationship between Client and Consultant, nor shall anything in this Agreement be deemed to create an agency relationship between Consultant and Client. Neither Consultant nor Client shall be or become liable or bound by any representation, act or omission whatsoever of the other. If any portion, term or provision of this Agreement shall be held illegal void or ineffective under, or in conflict with, any applicable law, the validity of the remaining provisions shall not be affected thereby, and a substitute provision will be negotiated to preserve as near as possible the original.intent of the Agreement.
考点1:“hold harmless and defend”建议译为“使……免受损害,并为其进行抗辩”,漏“defend”判错,体现组合表达 考点2:“any loss or damage”并列时须译作“任何损失或损害”,不可合并成“损失” 考点3:“servants”应译为“服务人员、雇员、代理人或承包商”,不宜漏译,应与“employees, agents, contractors”并列处理 考点4:“shall not exceed…”建议直译为“不得超过……”,避免“该等赔偿不应超过”式弱化表达 考点5:“provided, however, that…”建议译为“但前提是”或“但条件是”,避免“但该等赔偿不应超过……”模糊主语 考点6:“as a result of the offending Party's fault, omission or activities”建议译为“因过错方/侵权方的过错、不作为或行为造成”,避免简化为“由于该方……导致……”;“offending Party”优先译为“过错方/侵权方”,不可译作“违约方” 考点7:“promptly notifies”应译为“及时通知”;“gives... the right to control”建议译为“赋予……控制权”;“gives full cooperation”应译为“给予充分合作”;“complies with Consultant’s direction”宜译为“遵从……指示”。四项并列内容建议保留**“(i)(ii)(iii)(iv)”等编号结构,增强条款层次感和逻辑清晰度,避免散列句式弱化义务承担; 考点8:“waives all rights… howsoever arising”推荐译为“不可撤销地放弃……一切权利,且无论以何种方式/原因产生”,保持法律术语严谨性 考点9:“forever releases and discharges…”建议译为“永远免除和解除”,保留法律解除义务和权利的双重含义 考点10:“any and all claims, suits, losses…”建议译为“任何及所有索赔、诉讼、损失……”,不可省略“any and all”结构强调义,“any and all …”保留为“任何及所有……”,只译“任何”视为次优 考点11:“alleged or actual breach or failure…”译为“被指控的或实际的违反或未履行……”,需体现“alleged”之未决状态 考点12:“grossedup”应译为“返还计还原”或“税前补偿”,体现“税后净得”概念 考点13:“tax deduction, withholding or payment”建议按术语对应译为“税款扣除、预扣或缴纳” 考点14:“constitute... an agent, legal representative, subsidiary, joint venturer, partner, employer, or employee”应完整译为“构成一方为另一方的代理人、法定代表人、子公司、合营方、合作伙伴、雇主或雇员”;“shall not be construed to...”宜译为“不得被解释为……”。此类身份排除性表达应使用“不构成……亦不得解释为……”双重否定结构,增强法律约束力; 考点15:“employment taxes”推荐译为“就业税”,不应漏译为“用工相关税” 考点16:“imposts, levies, rates...”等多类税收推荐采用“征收、课征、税率”等合适对应术语,避免合并简化为“税收”一词 考点17:“shall not be or become liable or bound by any representation, act or omission whatsoever of the other”两层后果并列保留“不承担责任,亦不受其约束”,且不得遗漏“omission”“whatsoever” 考点18:“a substitute provision will be negotiated to preserve…”建议译为“由双方协商达成一项尽可能接近原意的替代条款”,避免模糊为“达成新的条款” 考点19:“reimburse... or provide... exemption certificates”应译为“(客户应)偿付(顾问已缴纳的税款),或(在顾问缴纳前向其)提供(有效的)免税证明”,结构表示二选一义务,。避免中译为“若……则……或……”等逻辑模糊句式,应准确呈现“偿付或提供证明”二选一义务关系 考点20:“negligence, breach of duty or other default or wrongdoing”须全译为“过失、违反义务或其他违约或不当行为” 考点21:“in Consultant’s sole judgment”必须体现“单方/独自判断”的权利 考点22:“patent issued as of the date of this Agreement”译为“截至本协议签署之日已授权的专利”,不可译“已公告” 考点23:“Party A, its affiliates, subsidiaries and related companies”三类主体分别列出,不可统称“关联公司”
46f4b
垂类场景
法律
104
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Another set of unresolved problems of interpretation concerns the traditional society – its social structure, the administrative institutions and ideology of its government and its economic growth. One widely held assumption is that this traditional society had created such an effective and balanced structure of ideas and practices that an innovative response to Western contact was difficult. In this view China’s ‘maturity’ was evidenced in her stability, her capacity to maintain a steady state almost like the homeostasis of physiology. To put it another way, the accumulation of vested interests was so great as to inhibit change. The result was a tremendous inertia or persistence in established ways, a tendency to change only within tradition. This notion of China’s inertial momentum and non-responsiveness is supported also by the concept of the autonomy of culture – that Chinese ways were different, mutually reinforcing, and therefore resistant to outside stimuli. Broad conceptions of this sort may of course merely substitute for thought. No doubt they display the degree of our present ignorance. Yet if history is to give understanding to non-historians, it must use general ideas. The essential prerequisite for generalizing is a factual picture of the Chinese state and society, including its general conditions and institutional practices, about 1800. At this time China’s population was over 300 million, almost double that of Europe including Russia, and it is safe to say that her home market and domestic trade were also far greater than those of Europe. The inertial momentum of China's social order was evident in the early nineteenth century at every level – among the common people in the peasant villages and market towns, among the dominant big families of the local elite or ‘gentry’, in the layers of imperial administration rising from the local magistracies all the way up to the court at Peking, and in the monarchy at the top of the human scene. This Chinese world (t'ien-hsia, ‘all under heaven’) was considered to be, and to a large extent was, remarkably unified, homogeneous and long-lasting. The unity of the empire had been the first great achievement of Chinese civilization and it remained a major concern, for unity meant peace. And yet the mere size and diversity of the empire often made for disunity. The eighteen provinces were divided by nature into a number of discrete and clearly marked regions, each comparatively self-sufficient. The central Taiyuan plain and Fen river valley of Shansi province, for example, were bounded on two sides by mountains and on two sides by the Yellow River. The great irrigated basin of Szechwan was ringed by mountains and communicated with the rest of China mainly through the Yangtze gorges. Yunnan province in the south-west was a plateau not in easy touch with the rest of the country. The great rice baskets of the Canton delta, the Yangtze delta, and Hunan and Hupei provinces each provided a base for local power. South Manchuria, as foreigners have called it in the twentieth century, or in Chinese terms of 1800 Liao-tung, was another power base where the Ch’ing dynasty of the Manchus had built up its strength before taking over south China with the Wall in 1644. Moreover, China stretched so far from north to south that the climatic difference created contrasting ways of life in the two regions. In the south and south-east, the heavy monsoon rainfall in summer facilitated double-cropping of rice. In contrast, the sparseness and unpredictability of rainfall on the north-western Great Wall frontier left people there constantly in danger of famine. The dry farmers in the north could live in houses of tamped earth walls or simple sun-baked brick with thatched roofs, while the wet farmers of the south had to use kiln-baked brick and tile roofs. They also wore straw sandals or clogs instead of cloth shoes, and broad plaited hats against rain and sun rather than the fur-covered caps of the north with ear flaps against winter cold. Much transport in the south was on waterways or, alternatively, on stone-paved paths not adequate for wheeled vehicles. Carrying poles, barrows and donkeys were universal, but the typical north China transport was by two-wheeled cart along dusty roads, frequently sunk several feet down into the loess soil scoured out by the wind. Most striking was the scenic contrast between the broad north China plain, dotted by walled villages that could defend themselves against cavalry raiders, and the typical hill country of south China where cavalry were useless and farmsteads might be more widely scattered in smaller units amid a lush cover of trees and vegetation. Since irrigated rice culture was far more productive than dry cereal farming, south China had a higher per capita food supply and more landlordism and tenancy among farmers. The gradual spread of Chinese civilization, with its characteristic features of intensive agriculture, tightly-organized family life, and bureaucratic administration, had given an underlying homogeneity to the whole country north and south, east and west. Perhaps this homogeneity was greater in the minds of the ruling elite than a sociologist would have found it to be in fact. Nevertheless, it was generally assumed. Like political unity, cultural homogeneity was one of China’s great social myths, demonstrating the universality of the Confucian way of life. Consequently, regional differences and the forms of localism have not yet been much studied, for it had always been and still is the fashion to discuss the vast land of China as a unit. The sense of unity and homogeneity had been fostered by the extraordinary continuity of the Chinese way of life, which came down undisrupted from neolithic times before the dawn of history. Hoe agriculture by family groups in settled villages had emerged by the fifth millennium BC in the Wei valley near the great bend of the Yellow River (as at the site of Pan-p’o outside Sian). Despite occasional invasions of warrior-rulers, Chinese village life had evolved steadily from that time with a continuity seemingly unbroken by sudden changes, either social or technological. The maintenance of peace and order among the village communities had been the special concern of China’s equally ancient ruling class. Under successive dynasties it had gradually created complex institutions of bureaucratic government. Until after 1800 the agrarian-bureaucratic empire of China thus preserved a social order more ancient than, and very different from, the commercial-military society of Europe. The individual’s prowess and aggressiveness, including his use of violence, had not been fostered in China’s agrarian communities as much as in the seafaring, warfare, exploration and overseas emigration of the Europeans. By 1800 we may assume that the ordinary peasants, who formed at least four-fifths of the population, were cultured individuals in the sense that they were well-schooled in the bonds of kinship, the duties of status, and the forms of politeness and social deportment, but generally illiterate or only semi-literate. Consequently their lives were less devoted to Confucian rationalism than to the lore, superstitions and Taoist-Buddhist religious observances of the folk culture. As farmers, most of them lived close to nature. They were accustomed to nature’s beauties but also suffered from epidemic diseases, for example, the prevalence of eye and skin ailments and intestinal parasites. As commoners, they were well aware of the ruling elite and its prerogatives, but they saw little of it directly and were chiefly absorbed in their own village-and-market-centre community. The usual village, perhaps one hundred households, was below the market level and not self-sufficient. Its real community centred on the market town, which was, of course, within walking distance not more than two or three miles away, so that a family member might go there and return on a periodic market day. The schedule of markets in one town, say on the third, sixth and eighth days of the ten-day cycle, would be integrated with the schedules of the adjacent towns, which might have their markets on the second, fourth, seventh and ninth days, or the third, fifth, eighth and tenth days of the cycle. In this way itinerant pedlars and merchants who operated out of a still larger market centre, could make the rounds of the market towns in its region. The lowest level or standard market town was usually surrounded by a dozen or eighteen villages, totalling perhaps fifteen hundred households or seven thousand people. From one of these village households an able-bodied male through the years might visit the market town a thousand times, and thus in its tea houses, its local temple or its occasional fairs and holiday celebrations have an opportunity to meet a large proportion of his market community. The basis of this community was not only economic – to exchange surplus farm or handicraft products so as to secure paper, iron implements, ceramic-ware or other imports – but also social. Since many villages were dominated by one kinship group, the rule of exogamous marriage led families to seek brides in other villages, usually through market town matchmakers. A secret society lodge, if present, would normally be centred in the market town, and there also the peasant would meet any local members of the ruling class or representatives of officialdom. In this peasant society, the individual depended upon his own kinship group for his livelihood and for the security provided in modern societies through insurance, as well as for education, recreation and major social contact. From infancy he was trained to observe the proper familial relationships, especially filial piety. The three bonds (san-kang) of the classical teaching were authoritarian, enjoining obedience to parents, subordination of wives to husbands, and loyalty of subject to ruler. But the hierarchy of status within the nuclear family was only part of the kinship network that extended out into the common descent group or lineage with which most families were connected.
Another set of unresolved problems of interpretation concerns the traditional society – its social structure, the administrative institutions and ideology of its government and its economic growth. One widely held assumption is that this traditional society had created such an effective and balanced structure of ideas and practices that an innovative response to Western contact was difficult. In this view China’s ‘maturity’ was evidenced in her stability, her capacity to maintain a steady state almost like the homeostasis of physiology. To put it another way, the accumulation of vested interests was so great as to inhibit change. The result was a tremendous inertia or persistence in established ways, a tendency to change only within tradition. This notion of China’s inertial momentum and non-responsiveness is supported also by the concept of the autonomy of culture – that Chinese ways were different, mutually reinforcing, and therefore resistant to outside stimuli. Broad conceptions of this sort may of course merely substitute for thought. No doubt they display the degree of our present ignorance. Yet if history is to give understanding to non-historians, it must use general ideas. The essential prerequisite for generalizing is a factual picture of the Chinese state and society, including its general conditions and institutional practices, about 1800. At this time China’s population was over 300 million, almost double that of Europe including Russia, and it is safe to say that her home market and domestic trade were also far greater than those of Europe. The inertial momentum of China's social order was evident in the early nineteenth century at every level – among the common people in the peasant villages and market towns, among the dominant big families of the local elite or ‘gentry’, in the layers of imperial administration rising from the local magistracies all the way up to the court at Peking, and in the monarchy at the top of the human scene. This Chinese world (t'ien-hsia, ‘all under heaven’) was considered to be, and to a large extent was, remarkably unified, homogeneous and long-lasting. The unity of the empire had been the first great achievement of Chinese civilization and it remained a major concern, for unity meant peace. And yet the mere size and diversity of the empire often made for disunity. The eighteen provinces were divided by nature into a number of discrete and clearly marked regions, each comparatively self-sufficient. The central Taiyuan plain and Fen river valley of Shansi province, for example, were bounded on two sides by mountains and on two sides by the Yellow River. The great irrigated basin of Szechwan was ringed by mountains and communicated with the rest of China mainly through the Yangtze gorges. Yunnan province in the south-west was a plateau not in easy touch with the rest of the country. The great rice baskets of the Canton delta, the Yangtze delta, and Hunan and Hupei provinces each provided a base for local power. South Manchuria, as foreigners have called it in the twentieth century, or in Chinese terms of 1800 Liao-tung, was another power base where the Ch’ing dynasty of the Manchus had built up its strength before taking over south China with the Wall in 1644. Moreover, China stretched so far from north to south that the climatic difference created contrasting ways of life in the two regions. In the south and south-east, the heavy monsoon rainfall in summer facilitated double-cropping of rice. In contrast, the sparseness and unpredictability of rainfall on the north-western Great Wall frontier left people there constantly in danger of famine. The dry farmers in the north could live in houses of tamped earth walls or simple sun-baked brick with thatched roofs, while the wet farmers of the south had to use kiln-baked brick and tile roofs. They also wore straw sandals or clogs instead of cloth shoes, and broad plaited hats against rain and sun rather than the fur-covered caps of the north with ear flaps against winter cold. Much transport in the south was on waterways or, alternatively, on stone-paved paths not adequate for wheeled vehicles. Carrying poles, barrows and donkeys were universal, but the typical north China transport was by two-wheeled cart along dusty roads, frequently sunk several feet down into the loess soil scoured out by the wind. Most striking was the scenic contrast between the broad north China plain, dotted by walled villages that could defend themselves against cavalry raiders, and the typical hill country of south China where cavalry were useless and farmsteads might be more widely scattered in smaller units amid a lush cover of trees and vegetation. Since irrigated rice culture was far more productive than dry cereal farming, south China had a higher per capita food supply and more landlordism and tenancy among farmers. The gradual spread of Chinese civilization, with its characteristic features of intensive agriculture, tightly-organized family life, and bureaucratic administration, had given an underlying homogeneity to the whole country north and south, east and west. Perhaps this homogeneity was greater in the minds of the ruling elite than a sociologist would have found it to be in fact. Nevertheless, it was generally assumed. Like political unity, cultural homogeneity was one of China’s great social myths, demonstrating the universality of the Confucian way of life. Consequently, regional differences and the forms of localism have not yet been much studied, for it had always been and still is the fashion to discuss the vast land of China as a unit. The sense of unity and homogeneity had been fostered by the extraordinary continuity of the Chinese way of life, which came down undisrupted from neolithic times before the dawn of history. Hoe agriculture by family groups in settled villages had emerged by the fifth millennium BC in the Wei valley near the great bend of the Yellow River (as at the site of Pan-p’o outside Sian). Despite occasional invasions of warrior-rulers, Chinese village life had evolved steadily from that time with a continuity seemingly unbroken by sudden changes, either social or technological. The maintenance of peace and order among the village communities had been the special concern of China’s equally ancient ruling class. Under successive dynasties it had gradually created complex institutions of bureaucratic government. Until after 1800 the agrarian-bureaucratic empire of China thus preserved a social order more ancient than, and very different from, the commercial-military society of Europe. The individual’s prowess and aggressiveness, including his use of violence, had not been fostered in China’s agrarian communities as much as in the seafaring, warfare, exploration and overseas emigration of the Europeans. By 1800 we may assume that the ordinary peasants, who formed at least four-fifths of the population, were cultured individuals in the sense that they were well-schooled in the bonds of kinship, the duties of status, and the forms of politeness and social deportment, but generally illiterate or only semi-literate. Consequently their lives were less devoted to Confucian rationalism than to the lore, superstitions and Taoist-Buddhist religious observances of the folk culture. As farmers, most of them lived close to nature. They were accustomed to nature’s beauties but also suffered from epidemic diseases, for example, the prevalence of eye and skin ailments and intestinal parasites. As commoners, they were well aware of the ruling elite and its prerogatives, but they saw little of it directly and were chiefly absorbed in their own village-and-market-centre community. The usual village, perhaps one hundred households, was below the market level and not self-sufficient. Its real community centred on the market town, which was, of course, within walking distance not more than two or three miles away, so that a family member might go there and return on a periodic market day. The schedule of markets in one town, say on the third, sixth and eighth days of the ten-day cycle, would be integrated with the schedules of the adjacent towns, which might have their markets on the second, fourth, seventh and ninth days, or the third, fifth, eighth and tenth days of the cycle. In this way itinerant pedlars and merchants who operated out of a still larger market centre, could make the rounds of the market towns in its region. The lowest level or standard market town was usually surrounded by a dozen or eighteen villages, totalling perhaps fifteen hundred households or seven thousand people. From one of these village households an able-bodied male through the years might visit the market town a thousand times, and thus in its tea houses, its local temple or its occasional fairs and holiday celebrations have an opportunity to meet a large proportion of his market community. The basis of this community was not only economic – to exchange surplus farm or handicraft products so as to secure paper, iron implements, ceramic-ware or other imports – but also social. Since many villages were dominated by one kinship group, the rule of exogamous marriage led families to seek brides in other villages, usually through market town matchmakers. A secret society lodge, if present, would normally be centred in the market town, and there also the peasant would meet any local members of the ruling class or representatives of officialdom. In this peasant society, the individual depended upon his own kinship group for his livelihood and for the security provided in modern societies through insurance, as well as for education, recreation and major social contact. From infancy he was trained to observe the proper familial relationships, especially filial piety. The three bonds (san-kang) of the classical teaching were authoritarian, enjoining obedience to parents, subordination of wives to husbands, and loyalty of subject to ruler. But the hierarchy of status within the nuclear family was only part of the kinship network that extended out into the common descent group or lineage with which most families were connected.
考点1:assumption 应译为 认定,主张,看法。不应译为“假设” 考点2:inertia or persistence 应译为“惯性,或称为持久性” 考点3:resistant应译为抗拒或对……有抵抗力,不应译为“抵制”或“抵御” 考点4:substitute for thought推荐译为“替代深入思考” 考点5:market town应译为“市镇” 考点6:the eighteen provinces were divided by nature…… 应译为十八省因自然地理条件被分割为 考点7: and Hunan and Hubei provinces 应译为湖广,因为两地在此处连在一起表述 考点8:power base 应译为政权或势力的根据地 考点9:taking over south of the Wall 应译为占领了长城以南的地区 考点10:intensive agriculture在历史学里应翻译为精耕农业,而不是集约化农业 考点11: Confucian way of life应译为“儒家之道”“儒家的典型特征”等,不应直译为“生活方式” 考点12:hoe agriculture应意译为耕种农业 考点13:village-and-market-centre community应译为以村和市镇为中心的社区 考点14:insurance 应译为生活保障,指现代社会提供的生活保障,并非保险
47978
学术论文
人文科学
170
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 跨境并购中的反垄断审查与国家安全审查协同机制 一、审查标准的冲突与协调 跨境并购的双重审查体系(反垄断审查与国家安全审查)常因标准差异引发合规困境。反垄断审查聚焦 “市场集中度”(HHI 指数)与 “竞争损害可能性”,而国家安全审查侧重 “关键领域控制权” 与 “供应链安全”,二者的核心冲突体现在三方面: 1. 审查范围的重叠与空白 根据《国务院反垄断委员会关于平台经济领域反垄断指南》,“数据集中” 可纳入反垄断审查;而《网络安全审查办法》将 “数据跨境” 列为国家安全审查重点。某跨国互联网企业收购中国社交平台时,因 “合并后市场份额达 35%” 触发反垄断审查,同时因 “掌握 1 亿用户生物信息” 被要求进行网络安全审查,导致同一交易面临两次审查,流程耗时延长 4 个月。反之,某些涉及 “非敏感技术但高市场份额” 的并购(如农用机械制造企业收购),因未落入国家安全审查范围,仅通过反垄断审查即完成交易,引发 “安全审查覆盖不全” 的争议。 2. 审查标准的实质性差异 反垄断审查中的 “相关市场界定”(如 “即时通讯服务市场”“云计算服务市场”)与国家安全审查中的 “关键领域划分”(如 “关键信息基础设施”“核心数据”)缺乏统一标准。例如,某半导体企业收购案中,反垄断机构界定相关市场为 “28 纳米以下芯片制造”,认定 “合并后市场集中度未达垄断阈值”;而国家安全审查机构将其归为 “核心元器件领域”,以 “可能影响供应链安全” 为由否决交易,标准冲突导致企业合规成本倍增。 3. 审查期限的衔接问题 《反垄断法》规定经营者集中审查期限为 30 日(初步审查)+90 日(进一步审查),特殊情况可延长 60 日;而《国家安全审查办法》未明确审查期限,实践中平均耗时 6-12 个月。某能源企业跨境并购案中,反垄断审查已通过,但国家安全审查持续 14 个月,导致交易因超过 “签约后 18 个月生效” 的条款而失效,企业损失定金 2 亿元。 二、审查豁免与例外情形的适用争议 两类审查均设有豁免条款,但适用条件的模糊性引发实践分歧: 1. “小额交易豁免” 的认定分歧 反垄断审查对 “交易额低于 4 亿元且各方市场份额低于 15%” 的交易可豁免审查,但国家安全审查无类似规定。某外资企业收购中国区域性连锁超市(交易额 3.8 亿元,市场份额 12%),虽符合反垄断豁免条件,却因 “涉及生鲜供应链数据” 被纳入国家安全审查,企业质疑 “小额交易不应触发安全审查”,但现行法规未明确排除。 2. “危机并购例外” 的适用空白 当目标企业面临破产时,反垄断审查可适用 “濒临破产企业抗辩”(如《横向合并指南》规定的 “若无并购则企业将退出市场”),但国家安全审查未设立类似例外。2023 年某国产芯片设计企业因资金链断裂拟被外资收购,反垄断审查以 “有助于维持市场竞争” 批准,而国家安全审查以 “技术外流风险” 否决,最终企业破产清算,引发 “安全审查是否应兼顾市场效率” 的讨论。 3. “国有资本并购” 的审查差异 国有企业跨境并购时,反垄断审查适用与民营企业相同的标准(如某央企收购海外能源公司因 “全球市场份额超 50%” 被附加剥离资产条件),但国家安全审查对 “国有资本控股” 的交易往往放宽标准,某国企收购境外港口企业时,因 “涉及关键基础设施” 本应从严审查,却以 “国有资本保障安全” 为由快速通过,被质疑 “审查标准双重化”。 三、数字化背景下的新型审查挑战 数据资产与新兴技术的跨境流动,使两类审查面临协同难题: 1. 数据并购的双重审查困境 企业通过 “资产收购” 获取数据时,若数据量未达 “经营者集中” 阈值(如收购仅含 10 万用户数据的小公司),可规避反垄断审查,但可能因 “数据敏感程度”(如健康数据)触发国家安全审查。某跨境医疗 APP 收购案中,收购标的用户量仅 8 万(未达反垄断申报标准),却因 “包含 HIV 感染者数据” 被要求进行网络安全审查,企业因未提前预判而导致交易搁置。 2. “扼杀式并购” 的识别盲区 大型科技企业收购初创公司以消灭潜在竞争对手(即 “扼杀式并购”),可能不违反反垄断法(因初创公司市场份额低),但可能涉及 “核心技术流失”。某海外巨头收购中国 AI 算法初创公司(市场份额 0.3%),反垄断审查未予禁止,而国家安全审查以 “算法可用于军事仿真” 为由否决,凸显 “非竞争因素主导审查” 的新趋势。 3. 审查信息共享的障碍 反垄断审查机构需获取 “市场份额数据”“竞争影响分析”,国家安全审查机构需掌握 “技术敏感程度”“供应链依赖度”,但两类信息分属不同部门管理,且涉及商业秘密与国家秘密的界限。某汽车行业并购案中,反垄断机构要求共享国家安全审查中的 “技术评估报告” 被拒,导致无法全面评估 “合并对新能源汽车技术竞争的影响”。 四、协同审查机制的完善路径 实践中已探索三类协同模式,但仍需优化: 1. “一案双报” 的流程整合 上海试点 “反垄断与国家安全审查联办窗口”,企业提交一套材料即可同时启动两类审查,审查期限按 “就长不就短” 原则(如国家安全审查耗时较长则以其为准)。某集成电路企业通过该窗口完成审查,总耗时较分开申报缩短 50%,但 “材料复用率” 仅 60%(因部分信息需分别适配两类审查要求)。 2. 审查标准的衔接清单 商务部发布《跨境并购敏感领域清单》,明确 “同时触发两类审查的领域”(如人工智能、生物医药)及 “优先适用安全审查的情形”(如涉及军事配套产业)。某生物制药企业并购案依据清单直接进入联合审查,避免了 “重复论证”,但清单未涵盖 “数据跨境 + 市场垄断” 的复合情形。 3. 争议解决的会商机制 设立跨部门会商小组,对标准冲突案件进行联合评审。2024 年某无人机企业并购案中,反垄断机构认为 “市场集中度未超标”,安全审查机构担忧 “飞控技术外流”,经会商后附加 “技术授权仅限民用领域” 的条件,促成交易通过,为类似案件提供了参考。
跨境并购中的反垄断审查与国家安全审查协同机制 一、审查标准的冲突与协调 跨境并购的双重审查体系(反垄断审查与国家安全审查)常因标准差异引发合规困境。反垄断审查聚焦 “市场集中度”(HHI 指数)与 “竞争损害可能性”,而国家安全审查侧重 “关键领域控制权” 与 “供应链安全”,二者的核心冲突体现在三方面: 1. 审查范围的重叠与空白 根据《国务院反垄断委员会关于平台经济领域反垄断指南》,“数据集中” 可纳入反垄断审查;而《网络安全审查办法》将 “数据跨境” 列为国家安全审查重点。某跨国互联网企业收购中国社交平台时,因 “合并后市场份额达 35%” 触发反垄断审查,同时因 “掌握 1 亿用户生物信息” 被要求进行网络安全审查,导致同一交易面临两次审查,流程耗时延长 4 个月。反之,某些涉及 “非敏感技术但高市场份额” 的并购(如农用机械制造企业收购),因未落入国家安全审查范围,仅通过反垄断审查即完成交易,引发 “安全审查覆盖不全” 的争议。 2. 审查标准的实质性差异 反垄断审查中的 “相关市场界定”(如 “即时通讯服务市场”“云计算服务市场”)与国家安全审查中的 “关键领域划分”(如 “关键信息基础设施”“核心数据”)缺乏统一标准。例如,某半导体企业收购案中,反垄断机构界定相关市场为 “28 纳米以下芯片制造”,认定 “合并后市场集中度未达垄断阈值”;而国家安全审查机构将其归为 “核心元器件领域”,以 “可能影响供应链安全” 为由否决交易,标准冲突导致企业合规成本倍增。 3. 审查期限的衔接问题 《反垄断法》规定经营者集中审查期限为 30 日(初步审查)+90 日(进一步审查),特殊情况可延长 60 日;而《国家安全审查办法》未明确审查期限,实践中平均耗时 6-12 个月。某能源企业跨境并购案中,反垄断审查已通过,但国家安全审查持续 14 个月,导致交易因超过 “签约后 18 个月生效” 的条款而失效,企业损失定金 2 亿元。 二、审查豁免与例外情形的适用争议 两类审查均设有豁免条款,但适用条件的模糊性引发实践分歧: 1. “小额交易豁免” 的认定分歧 反垄断审查对 “交易额低于 4 亿元且各方市场份额低于 15%” 的交易可豁免审查,但国家安全审查无类似规定。某外资企业收购中国区域性连锁超市(交易额 3.8 亿元,市场份额 12%),虽符合反垄断豁免条件,却因 “涉及生鲜供应链数据” 被纳入国家安全审查,企业质疑 “小额交易不应触发安全审查”,但现行法规未明确排除。 2. “危机并购例外” 的适用空白 当目标企业面临破产时,反垄断审查可适用 “濒临破产企业抗辩”(如《横向合并指南》规定的 “若无并购则企业将退出市场”),但国家安全审查未设立类似例外。2023 年某国产芯片设计企业因资金链断裂拟被外资收购,反垄断审查以 “有助于维持市场竞争” 批准,而国家安全审查以 “技术外流风险” 否决,最终企业破产清算,引发 “安全审查是否应兼顾市场效率” 的讨论。 3. “国有资本并购” 的审查差异 国有企业跨境并购时,反垄断审查适用与民营企业相同的标准(如某央企收购海外能源公司因 “全球市场份额超 50%” 被附加剥离资产条件),但国家安全审查对 “国有资本控股” 的交易往往放宽标准,某国企收购境外港口企业时,因 “涉及关键基础设施” 本应从严审查,却以 “国有资本保障安全” 为由快速通过,被质疑 “审查标准双重化”。 三、数字化背景下的新型审查挑战 数据资产与新兴技术的跨境流动,使两类审查面临协同难题: 1. 数据并购的双重审查困境 企业通过 “资产收购” 获取数据时,若数据量未达 “经营者集中” 阈值(如收购仅含 10 万用户数据的小公司),可规避反垄断审查,但可能因 “数据敏感程度”(如健康数据)触发国家安全审查。某跨境医疗 APP 收购案中,收购标的用户量仅 8 万(未达反垄断申报标准),却因 “包含 HIV 感染者数据” 被要求进行网络安全审查,企业因未提前预判而导致交易搁置。 2. “扼杀式并购” 的识别盲区 大型科技企业收购初创公司以消灭潜在竞争对手(即 “扼杀式并购”),可能不违反反垄断法(因初创公司市场份额低),但可能涉及 “核心技术流失”。某海外巨头收购中国 AI 算法初创公司(市场份额 0.3%),反垄断审查未予禁止,而国家安全审查以 “算法可用于军事仿真” 为由否决,凸显 “非竞争因素主导审查” 的新趋势。 3. 审查信息共享的障碍 反垄断审查机构需获取 “市场份额数据”“竞争影响分析”,国家安全审查机构需掌握 “技术敏感程度”“供应链依赖度”,但两类信息分属不同部门管理,且涉及商业秘密与国家秘密的界限。某汽车行业并购案中,反垄断机构要求共享国家安全审查中的 “技术评估报告” 被拒,导致无法全面评估 “合并对新能源汽车技术竞争的影响”。 四、协同审查机制的完善路径 实践中已探索三类协同模式,但仍需优化: 1. “一案双报” 的流程整合 上海试点 “反垄断与国家安全审查联办窗口”,企业提交一套材料即可同时启动两类审查,审查期限按 “就长不就短” 原则(如国家安全审查耗时较长则以其为准)。某集成电路企业通过该窗口完成审查,总耗时较分开申报缩短 50%,但 “材料复用率” 仅 60%(因部分信息需分别适配两类审查要求)。 2. 审查标准的衔接清单 商务部发布《跨境并购敏感领域清单》,明确 “同时触发两类审查的领域”(如人工智能、生物医药)及 “优先适用安全审查的情形”(如涉及军事配套产业)。某生物制药企业并购案依据清单直接进入联合审查,避免了 “重复论证”,但清单未涵盖 “数据跨境 + 市场垄断” 的复合情形。 3. 争议解决的会商机制 设立跨部门会商小组,对标准冲突案件进行联合评审。2024 年某无人机企业并购案中,反垄断机构认为 “市场集中度未超标”,安全审查机构担忧 “飞控技术外流”,经会商后附加 “技术授权仅限民用领域” 的条件,促成交易通过,为类似案件提供了参考。
考点 1:“《国务院反垄断委员会关于平台经济领域反垄断指南》” 必须译为 “Anti-monopoly Guidelines of the Anti-monopoly Commission of the State Council for the Field of Platform Economy”,固定译法 考点 2: “《网络安全审查办法” 必须译为 “The Measures for Cybersecurity Review”,固定译法 考点 3:“相关市场界定” 推荐译为 “relevant market definition” 考点 4:“《反垄断法》” 必须译为 “The Anti-Monopoly Law ”,固定译法 考点 5:“经营者集中” 推荐译为 “concentration of undertakings”,不能译为 “enterprise merger”,因为此处表示的不是“兼并”的意思 考点 6:“HHI ”必须译为“Herfindahl-Hirschman Index (HHI)”,固定译法,且首次出现应该用全称,之后再用简称 考点 7:“数据集中” 必须译为“data aggregation”。不能译为 “data concentration”,固定译法 考点 8:“技术外流风险” 推荐译为 “risk of technology outflow”
481f7
学术论文
社会科学
164
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 这几天心里颇不宁静。今晚在院子里坐着乘凉,忽然想起日日走过的荷塘,在这满月的光里,总该另有一番样子吧。月亮渐渐地升高了,墙外马路上孩子们的欢笑,已经听不见了;妻在屋里拍着闰儿,迷迷糊糊地哼着眠歌。我悄悄地披了大衫,带上门出去。   沿着荷塘,是一条曲折的小煤屑路。这是一条幽僻的路;白天也少人走,夜晚更加寂寞。荷塘四面,长着许多树,蓊蓊郁郁的。路的一旁,是些杨柳,和一些不知道名字的树。没有月光的晚上,这路上阴森森的,有些怕人。今晚却很好,虽然月光也还是淡淡的。   路上只我一个人,背着手踱着。这一片天地好像是我的;我也像超出了平常的自己,到了另一世界里。我爱热闹,也爱冷静;爱群居,也爱独处。像今晚上,一个人在这苍茫的月下,什么都可以想,什么都可以不想,便觉是个自由的人。白天里一定要做的事,一定要说的话,现在都可不理。这是独处的妙处,我且受用这无边的荷香月色好了。   曲曲折折的荷塘上面,弥望的是田田的叶子。叶子出水很高,像亭亭的舞女的裙。层层的叶子中间,零星地点缀着些白花,有袅娜地开着的,有羞涩地打着朵儿的;正如一粒粒的明珠,又如碧天里的星星,又如刚出浴的美人。微风过处,送来缕缕清香,仿佛远处高楼上渺茫的歌声似的。这时候叶子与花也有一丝的颤动,像闪电般,霎时传过荷塘的那边去了。叶子本是肩并肩密密地挨着,这便宛然有了一道凝碧的波痕。叶子底下是脉脉的流水,遮住了,不能见一些颜色;而叶子却更见风致了。   月光如流水一般,静静地泻在这一片叶子和花上。薄薄的青雾浮起在荷塘里。叶子和花仿佛在牛乳中洗过一样;又像笼着轻纱的梦。虽然是满月,天上却有一层淡淡的云,所以不能朗照;但我以为这恰是到了好处——酣眠固不可少,小睡也别有风味的。月光是隔了树照过来的,高处丛生的灌木,落下参差的斑驳的黑影,峭楞楞如鬼一般;弯弯的杨柳的稀疏的倩影,却又像是画在荷叶上。塘中的月色并不均匀;但光与影有着和谐的旋律,如梵婀玲上奏着的名曲。   荷塘的四面,远远近近,高高低低都是树,而杨柳最多。这些树将一片荷塘重重围住;只在小路一旁,漏着几段空隙,像是特为月光留下的。树色一例是阴阴的,乍看像一团烟雾;但杨柳的丰姿,便在烟雾里也辨得出。树梢上隐隐约约的是一带远山,只有些大意罢了。树缝里也漏着一两点路灯光,没精打采的,是渴睡人的眼。这时候最热闹的,要数树上的蝉声与水里的蛙声;但热闹是它们的,我什么也没有。   忽然想起采莲的事情来了。采莲是江南的旧俗,似乎很早就有,而六朝时为盛;从诗歌里可以约略知道。采莲的是少年的女子,她们是荡着小船,唱着艳歌去的。采莲人不用说很多,还有看采莲的人。那是一个热闹的季节,也是一个风流的季节。梁元帝《采莲赋》里说得好:   于是妖童媛女,荡舟心许;鷁首徐回,兼传羽杯;欋将移而藻挂,船欲动而萍开。尔其纤腰束素,迁延顾步;夏始春余,叶嫩花初,恐沾裳而浅笑,畏倾船而敛裾。   可见当时嬉游的光景了。这真是有趣的事,可惜我们现在早已无福消受了。   于是又记起《西洲曲》里的句子:   采莲南塘秋,莲花过人头;低头弄莲子,莲子清如水。今晚若有采莲人,这儿的莲花也算得“过人头”了;只不见一些流水的影子,是不行的。这令我到底惦着江南了。——这样想着,猛一抬头,不觉已是自己的门前;轻轻地推门进去,什么声息也没有,妻已睡熟好久了。
这几天心里颇不宁静。今晚在院子里坐着乘凉,忽然想起日日走过的荷塘,在这满月的光里,总该另有一番样子吧。月亮渐渐地升高了,墙外马路上孩子们的欢笑,已经听不见了;妻在屋里拍着闰儿,迷迷糊糊地哼着眠歌。我悄悄地披了大衫,带上门出去。   沿着荷塘,是一条曲折的小煤屑路。这是一条幽僻的路;白天也少人走,夜晚更加寂寞。荷塘四面,长着许多树,蓊蓊郁郁的。路的一旁,是些杨柳,和一些不知道名字的树。没有月光的晚上,这路上阴森森的,有些怕人。今晚却很好,虽然月光也还是淡淡的。   路上只我一个人,背着手踱着。这一片天地好像是我的;我也像超出了平常的自己,到了另一世界里。我爱热闹,也爱冷静;爱群居,也爱独处。像今晚上,一个人在这苍茫的月下,什么都可以想,什么都可以不想,便觉是个自由的人。白天里一定要做的事,一定要说的话,现在都可不理。这是独处的妙处,我且受用这无边的荷香月色好了。   曲曲折折的荷塘上面,弥望的是田田的叶子。叶子出水很高,像亭亭的舞女的裙。层层的叶子中间,零星地点缀着些白花,有袅娜地开着的,有羞涩地打着朵儿的;正如一粒粒的明珠,又如碧天里的星星,又如刚出浴的美人。微风过处,送来缕缕清香,仿佛远处高楼上渺茫的歌声似的。这时候叶子与花也有一丝的颤动,像闪电般,霎时传过荷塘的那边去了。叶子本是肩并肩密密地挨着,这便宛然有了一道凝碧的波痕。叶子底下是脉脉的流水,遮住了,不能见一些颜色;而叶子却更见风致了。   月光如流水一般,静静地泻在这一片叶子和花上。薄薄的青雾浮起在荷塘里。叶子和花仿佛在牛乳中洗过一样;又像笼着轻纱的梦。虽然是满月,天上却有一层淡淡的云,所以不能朗照;但我以为这恰是到了好处——酣眠固不可少,小睡也别有风味的。月光是隔了树照过来的,高处丛生的灌木,落下参差的斑驳的黑影,峭楞楞如鬼一般;弯弯的杨柳的稀疏的倩影,却又像是画在荷叶上。塘中的月色并不均匀;但光与影有着和谐的旋律,如梵婀玲上奏着的名曲。   荷塘的四面,远远近近,高高低低都是树,而杨柳最多。这些树将一片荷塘重重围住;只在小路一旁,漏着几段空隙,像是特为月光留下的。树色一例是阴阴的,乍看像一团烟雾;但杨柳的丰姿,便在烟雾里也辨得出。树梢上隐隐约约的是一带远山,只有些大意罢了。树缝里也漏着一两点路灯光,没精打采的,是渴睡人的眼。这时候最热闹的,要数树上的蝉声与水里的蛙声;但热闹是它们的,我什么也没有。   忽然想起采莲的事情来了。采莲是江南的旧俗,似乎很早就有,而六朝时为盛;从诗歌里可以约略知道。采莲的是少年的女子,她们是荡着小船,唱着艳歌去的。采莲人不用说很多,还有看采莲的人。那是一个热闹的季节,也是一个风流的季节。梁元帝《采莲赋》里说得好:   于是妖童媛女,荡舟心许;鷁首徐回,兼传羽杯;欋将移而藻挂,船欲动而萍开。尔其纤腰束素,迁延顾步;夏始春余,叶嫩花初,恐沾裳而浅笑,畏倾船而敛裾。   可见当时嬉游的光景了。这真是有趣的事,可惜我们现在早已无福消受了。   于是又记起《西洲曲》里的句子:   采莲南塘秋,莲花过人头;低头弄莲子,莲子清如水。今晚若有采莲人,这儿的莲花也算得“过人头”了;只不见一些流水的影子,是不行的。这令我到底惦着江南了。——这样想着,猛一抬头,不觉已是自己的门前;轻轻地推门进去,什么声息也没有,妻已睡熟好久了。
考点1:“颇不宁静”应该译为“ rather restless”,表示内心烦乱。 考点2:“乘凉”应译为“enjoying the cool”,是中国特有的夏日习惯。 考点3:“大衫”应译为“long robe”,中国特有服饰。 考点4:“踱着”应译为“strolling”,表示慢速、悠闲、带有沉思意味的散步。 考点5:“苍茫”应译为“vast and mistry ”,形容的是一种广阔、空旷、辽远而又有些模糊不清的景象。 考点6:“受用”应译为“indulge in”表示沉浸在这片风景之中。 考点7:“田田的叶子”应译为“an expanse of leaves”,表示一大片的荷叶。 考点8:“风致”应译为“attractive”表示吸引人的。 考点9:“梵炯铃”应译为“violin”,当时那个年代外来的音译词。 考点10:“风流的”应译为“romantic”,形容男女之间的浪漫爱情关系。 考点11:“梁元帝的”应译为“Emperor Yuan of the Liang Dynasty”。 考点12:“妖童媛女的”应译为“charming boys and lovely maidens”,指的是有魅力的男子和可爱的女子。 考点13:“惦着”应译为“miss”,表示想念。 考点14:“江南”应译为“south of the Yangtze River”,指明具体的地理位置。
49bb3
文学艺术
散文
51
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 关税缓和背景下,会议认为“主要经济指标表现良好”,强调“保持政策连续性稳定性”。日内瓦会谈后,美国对中国加征关税缓和,支撑了出口,在此带动下,二季度经济运行虽然较一季度小幅放缓,但整体平稳。会议认为“我国经济运行稳中有进”,“主要经济指标表现良好”,对于风险挑战的定调也由4月政治局会议中的“外部冲击影响加大”变为“依然面临不少风险挑战”。因此,政策方面延续了4月政治局会议中提出的“稳就业、稳企业、稳市场、稳预期”,更多强调“保持政策连续性稳定性”和“持续发力”,但也提出“增强灵活性预见性”和“适时加力”,我们认为后者更多是针对经济运行超预期走弱情形下的政策储备。 三季度或聚焦发挥已有政策效力,不排除四季度财政加码的可能性。会议指出,“宏观政策要持续发力、适时加力,要落实落细更加积极的财政政策,充分释放政策效应”,“适时加力”这一表述,或更多是针对经济运行超预期走弱情形下的政策储备。展望三季度,财政政策或聚焦发挥已出台的财政政策工具效力,如加强对政府投资项目的督导,以配合政府债券的加快发行和使用。展望全年,财政收入存在低于预算目标的可能,若四季度经济超预期走弱,或出台增量财政政策,如提前下达明年部分政府债额度或者通过特定机构利润上缴等方式,以补足年内财政收支缺口。 存量债务的风险化解工作持续推进。会议指出要“积极稳妥化解地方政府债务风险,严禁新增隐性债务”,相较4月政治局会议,新增“有力有序有效推进地方融资平台出清”的表述。去年以来,地方政府陆续制定了区域内城投退平台的时间表,伴随着隐性债务的偿还加快,退平台的城投名单也将继续扩围。据财政部统计[2],2024年的2万亿元置换债券已于今年上半年全部使用完毕,今年的2万亿元置换债券截至6月末已发行1.8万亿元,已使用1.44万亿元。不过城投退平台后需遵循市场化的融资方式,这意味着城投扩表的难度增加,对地方基建的托举相对有限。 降低融资成本未必需要调降政策利率,改革性措施与结构性货币政策工具或为重点。本次政治局会议继续强调“货币政策要保持流动性充裕,促进社会综合融资成本下行”,从最近几个月的货币政策态度来看,降低综合融资成本未必意味着需要继续调降政策利率,《金融时报》在此前的文章《社会融资成本下降成效显著》中提出“目前我国贷款利息成本已经很低,未来降低综合融资成本关键是要降低抵押担保费、中介服务费等非利息成本”[3]。此外,本次政治局会议继续强调“用好各项结构性货币政策工具”,央行对经济运行薄弱环节以及可能遭受冲击的风险环节提供针对性的支持,有可能成为接下来货币政策的重要抓手。 培育服务消费,保障改善民生。会议提出“深入实施提振消费专项行动”,要求“在扩大商品消费的同时,培育服务消费新的增长点”。当前的以旧换新政策侧重于扩大商品消费,而培育发展服务消费在2023年以来的政治局会议中被多次提及。参考以往政策表述,主要指向了体育休闲、文化旅游等文娱休闲领域,以及养老、育幼、家政等基本民生领域。会议还提出“在保障改善民生中扩大消费需求”,民生或成为促消费政策的重要抓手。例如近期出台的育儿补贴制度、推行免费学前教育、向中度以上失能老年人发放养老服务消费补贴等,重点关注“一老一小”,在保民生、促消费的同时,也有助于应对人口结构变化、提高增长潜力。 加力支持科技创新。本次会议指出,要“坚持以科技创新引领新质生产力发展,加快培育具有国际竞争力的新兴支柱产业,推动科技创新和产业创新深度融合发展”,总体延续了此前对新质生产力的表述。我们认为,对科技创新、新质生产力的支持是未来的中长期主线。政治局会议也指出,要用好结构性货币政策工具,“加力支持科技创新”。2025年4月,政治局会议提出要“创新推出债券市场的‘科技板’”。根据人民银行数据,截至2025年5月,已经有将近100家左右的机构在发行科技创新债券,金额超过了2500亿元[4]。除了债券支持工具外,对科技创新的信贷支持也在加大。截至2025年3月末,科技型中小企业贷款余额已超过3.3万亿元,同比增长24%,连续三年增速超过20%[5]。 纵深推进全国统一大市场建设,推动市场竞争秩序持续优化,是理解“反内卷”政策的关键。此前《反不正当竞争法》完成修订,为治理企业的无序竞争提供了法律依据。近日两部委推出《价格法修正草案(征求意见稿)》[6],明确提出规范市场价格秩序,并加强成本监审。“反内卷”继续沿着市场化、法治化的治理思路推进,光伏、水泥、煤炭和化工等重点行业的产能调控政策后续落地实施,政策效果值得持续观察。在制度层面上,“内卷”的成因既有市场失灵,也表现为部分地方政府行政干预[7],规范地方招商引资行为,也是完善市场竞争秩序的重要一环。
关税缓和背景下,会议认为“主要经济指标表现良好”,强调“保持政策连续性稳定性”。日内瓦会谈后,美国对中国加征关税缓和,支撑了出口,在此带动下,二季度经济运行虽然较一季度小幅放缓,但整体平稳。会议认为“我国经济运行稳中有进”,“主要经济指标表现良好”,对于风险挑战的定调也由4月政治局会议中的“外部冲击影响加大”变为“依然面临不少风险挑战”。因此,政策方面延续了4月政治局会议中提出的“稳就业、稳企业、稳市场、稳预期”,更多强调“保持政策连续性稳定性”和“持续发力”,但也提出“增强灵活性预见性”和“适时加力”,我们认为后者更多是针对经济运行超预期走弱情形下的政策储备。 三季度或聚焦发挥已有政策效力,不排除四季度财政加码的可能性。会议指出,“宏观政策要持续发力、适时加力,要落实落细更加积极的财政政策,充分释放政策效应”,“适时加力”这一表述,或更多是针对经济运行超预期走弱情形下的政策储备。展望三季度,财政政策或聚焦发挥已出台的财政政策工具效力,如加强对政府投资项目的督导,以配合政府债券的加快发行和使用。展望全年,财政收入存在低于预算目标的可能,若四季度经济超预期走弱,或出台增量财政政策,如提前下达明年部分政府债额度或者通过特定机构利润上缴等方式,以补足年内财政收支缺口。 存量债务的风险化解工作持续推进。会议指出要“积极稳妥化解地方政府债务风险,严禁新增隐性债务”,相较4月政治局会议,新增“有力有序有效推进地方融资平台出清”的表述。去年以来,地方政府陆续制定了区域内城投退平台的时间表,伴随着隐性债务的偿还加快,退平台的城投名单也将继续扩围。据财政部统计[2],2024年的2万亿元置换债券已于今年上半年全部使用完毕,今年的2万亿元置换债券截至6月末已发行1.8万亿元,已使用1.44万亿元。不过城投退平台后需遵循市场化的融资方式,这意味着城投扩表的难度增加,对地方基建的托举相对有限。 降低融资成本未必需要调降政策利率,改革性措施与结构性货币政策工具或为重点。本次政治局会议继续强调“货币政策要保持流动性充裕,促进社会综合融资成本下行”,从最近几个月的货币政策态度来看,降低综合融资成本未必意味着需要继续调降政策利率,《金融时报》在此前的文章《社会融资成本下降成效显著》中提出“目前我国贷款利息成本已经很低,未来降低综合融资成本关键是要降低抵押担保费、中介服务费等非利息成本”[3]。此外,本次政治局会议继续强调“用好各项结构性货币政策工具”,央行对经济运行薄弱环节以及可能遭受冲击的风险环节提供针对性的支持,有可能成为接下来货币政策的重要抓手。 培育服务消费,保障改善民生。会议提出“深入实施提振消费专项行动”,要求“在扩大商品消费的同时,培育服务消费新的增长点”。当前的以旧换新政策侧重于扩大商品消费,而培育发展服务消费在2023年以来的政治局会议中被多次提及。参考以往政策表述,主要指向了体育休闲、文化旅游等文娱休闲领域,以及养老、育幼、家政等基本民生领域。会议还提出“在保障改善民生中扩大消费需求”,民生或成为促消费政策的重要抓手。例如近期出台的育儿补贴制度、推行免费学前教育、向中度以上失能老年人发放养老服务消费补贴等,重点关注“一老一小”,在保民生、促消费的同时,也有助于应对人口结构变化、提高增长潜力。 加力支持科技创新。本次会议指出,要“坚持以科技创新引领新质生产力发展,加快培育具有国际竞争力的新兴支柱产业,推动科技创新和产业创新深度融合发展”,总体延续了此前对新质生产力的表述。我们认为,对科技创新、新质生产力的支持是未来的中长期主线。政治局会议也指出,要用好结构性货币政策工具,“加力支持科技创新”。2025年4月,政治局会议提出要“创新推出债券市场的‘科技板’”。根据人民银行数据,截至2025年5月,已经有将近100家左右的机构在发行科技创新债券,金额超过了2500亿元[4]。除了债券支持工具外,对科技创新的信贷支持也在加大。截至2025年3月末,科技型中小企业贷款余额已超过3.3万亿元,同比增长24%,连续三年增速超过20%[5]。 纵深推进全国统一大市场建设,推动市场竞争秩序持续优化,是理解“反内卷”政策的关键。此前《反不正当竞争法》完成修订,为治理企业的无序竞争提供了法律依据。近日两部委推出《价格法修正草案(征求意见稿)》[6],明确提出规范市场价格秩序,并加强成本监审。“反内卷”继续沿着市场化、法治化的治理思路推进,光伏、水泥、煤炭和化工等重点行业的产能调控政策后续落地实施,政策效果值得持续观察。在制度层面上,“内卷”的成因既有市场失灵,也表现为部分地方政府行政干预[7],规范地方招商引资行为,也是完善市场竞争秩序的重要一环。
考点1:“关税缓和”推荐译为“easing tariffs” 考点2:“稳中有进”推荐译为“steady progress","maintaining stability while making progress” 考点3:“定调”推荐译为“Set the tone” 考点4:“预见性”推荐译为“forward-looking” 考点5:“适时加力”推荐译为“well-timed measures” 考点6:“财政加码”推荐译为“increase fiscal stimulus” 考点7:“存量债务”推荐译为“outstanding debt” 考点8:“地方融资平台出清”推荐译为“clearing of local government financing vehicles” 考点9:“置换债券”推荐译为“Bond Swap” 考点10:“社会综合融资成本下行”推荐译为“a decline in overall social financing costs” 考点11:“提振消费专项行动”推荐译为“special action plan to boost consumption” 考点12:“科技板”必须译为“Sci-Tech Board ” 考点13:“全国统一大市场建设”必须译为“construction of a unified national market”
4ba49
新闻资讯
观点评论
138
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 一、 互动式电影概述 (一)互动式电影的概念 科学技术带动了传播媒介的转换以及影视与游戏形态的变革,互动式电影(Interactive Movie)是科学技术的不断发展发展产生的一种新型电影形式。互动式电影是传统电影结合不断发展的数字游戏与数字媒体技术所产生的概念,它是相交于传统电影和电子游戏之间的新型领域,相较于电子游戏,互动式电影更加倾向于影像质感与整体叙事;相较于传统电影,游戏大师Chris Crawford认为:“电影并没有交互性。电影以精心构思好的方式进行叙事,而互动叙事则是随着情节的不断发展逐步建立起来的。”[ ChrisCrawford. 游戏大师Chris Crawford谈互动叙事[M]. 人民邮电出版社, 2015.] 互动式电影让观众拥有更多自由选择的权利,对于观众来说,互动式电影突破了以往的界限,观众不仅仅是观众,观众也成为了叙事的参与者与创造者。与往常观众被动接受内容的传统电影不同,互动式电影在不同的叙事情节点设置不同的个性化选项,观众可以根据自己的意愿从提供的交互选项中选择一个延续剧情的发展,观众的差异化选择会进入不同的故事情节,最后形成多元的结局。 值得注意的是,观众选择的互动选项不一定是关键情节的关键选择,也有许多日常的交互,比如由2018年由奈飞(Netflix)出品的互动式电影·《黑镜:潘达斯奈基》中主角父亲早上问主角吃什么早餐,主角决定听什么音乐,这些细节都可以由观众个性选择,不同的选择会影响后续的剧情走向与结局。充满创造性和个性化的电影叙事,这是互动式电影带给观众一种崭新的介入性强的观影体验。现代电子游戏注重游戏操作和游戏机制,电子游戏玩家获得成就感的方式是通过自身的游戏操作技术突破一道道游戏关卡最终实现胜利,而互动式电影更注重的是通过交互机制增强故事的表现力,它没有获得胜利的条件也没有方向和路线,无论观众如何选择都不会导致失败,最终的结局都是由观众根据自身的意愿选择产生,都是拥有闭环叙事的结局。 黑镜:潘达斯奈基 (二)互动式电影的发展现状 一部63分钟的捷克短片叫做《一个男人与他的房子》于1967年出现在蒙特利尔世博会上,惊动了当时整个电影界,这是世界上第一部互动式电影,当时电影的播放渠道主要集中在影院放映,配备按钮和互动设备的高额成本和技术限制让影院望而却步,很快互动式电影便陷入沉寂之中。 随着科学技术的发展,电子计算机,电子移动设备的大量使用以及互联网的不断发展,人类进入了数字媒介时代,随着3D图形技术的逐渐成熟,电子游戏电影化形成的交互式电影游戏崭露头角,在2010年,游戏制作人大卫·凯奇的作品《暴雨》 打破了互动式电影游戏的沉寂。 超凡双生 二、互动式电影跨媒介叙事方式的探究 (一)超叙事互动:网状及分支叙事模式 传统的电影叙事是导演对电影情节整体安排下的一次线性叙事,因为这种叙事单线输出于观众的。我们不难发现其实传统电影中的所有线索往往都涵盖在一个固定的框架中,犹如交错的道路,导演只设置一条路线所通往的地点,可能线索伴随着时间或是空间进行线性推动。“超叙事互动” 这种多选择,多路线的网状及分支叙事模式是互动式电影最大特点之一,观众的每一次选择都在原有的故事情节上建立了新平行时空,互动者在现实世界中思考和抉择,互动者的每一个选择都会影响虚拟世界的剧情走向,从而推进故事的发展。如互动式电影《夜班》。 (二)开放式叙事:被弱化的间离效应 传统电影中观众观看电影的元素包括:剧情,画面,演员表演,特效动画,都是由电影导演与编剧们创造的一个“囚笼”,观众们从电影开始到结束始终困在这个盒子里,无论电影如何演绎都无法突破这个“囚笼”,这是传统电影的“封闭叙事”属性。而互动式电影通过融合电子游戏的交互模式让观众拥有了打破“囚笼”的能力,观众拥有了书写文本的能力,打破了创作者们打造的囚笼,“开放叙事”由此产生。 上文提及的《一个男人与他的房子》需要观众选择红绿按钮来决定不同的电影剧情走向。观众们在观影时不同的选择给与了观众们新奇感,但完成所有的选择后影片的结局却又让观众们食而无味,观众以为自己选择的不同按钮能够给电影带来不同的结局走向,但是他们的选择毫无意义,无论如何选择那个按钮走那条剧情线,影片的结局最终也只有一个,直到最后观众们还是被导演玩弄于股掌之中。通过这个例子我们不难发现其实最早的互动式电影中还是无法脱离传统电影的线性叙事思路,即所有线索往往都涵盖在一个固定的框架中,犹如交错的道路。 导演预先设置一个所有道路所通往的终点,观众则通过自身思考,选择不同的道路。观众设想的每一条道路都可以使整个观影过程完整。但是观众面对的剧情选择往往情感上仍有缺失,因为导演可以设定路线但不能对每一位观众面对的结局进行优化,线性电影的结局往往是封闭的,唯一的,一部分原因归结于当时电影的时长和容量限制了这种需求的达成。如今随着互联网与新媒体技术的发展,互动式电影真正让观众参与了叙事,实现了影游在叙事上的融合。观众真正加入叙事,并可以按照自身的意志对叙事进行创作,此时不同方向的开展路线拥有了不同的结局,传统的封闭式结构由此发生了改变。如互动式电影《深海》共有2条主要故事线,15条支线,将近60个交互点,高达25个结局;《黑镜:潘达斯奈基》主副线约5个小时,30多个交互点相互排列组合最终会产生5个各不相同的方向的15个结局;《夜班》整部电影更拥有高达180个可以选择的交互按钮,7个不同的结局。面对以往观众对传统电影唯一结局的情感缺失,互动式电影给予了观众重新再来的契机,错误可以被改正,惋惜的瞬间可以被重现。 现阶段互动式电影的开放式叙事更多的是对叙事文本的不确定性进行填充,可进行的交互选择往往还是在一定的范围之内,但已经产生了观众对叙事文本的能动性参与,弱化了间离效果。间离化效果是由戏剧家布莱希特提出的戏剧表演艺术的代表性观点,在电影艺术中普遍得到应用,它主要表现为利用旁白、标语、歌唱等形式提醒观众当前的故事为虚构而不是真实,并通过客观视角或第三人称视角对叙事进行修改和更正。[ 贝・布莱希特. 布莱希特论戏剧[M]. 中国戏剧出版社, 1990.]由于互动式电影更注重沉浸感和主体的情感体验,在互动式电影中制作者往往就会弱化间离效果,通过设置交互界面增强观众的沉浸感体验。为了达到间离效果的目的,互动式电影往往采用分章节分段落的方式让观众逐渐了解电影的世界观和虚构的故事框架,最终完成对故事叙事的认知,形成开放式叙事。
一、 互动式电影概述 (一)互动式电影的概念 科学技术带动了传播媒介的转换以及影视与游戏形态的变革,互动式电影(Interactive Movie)是科学技术的不断发展发展产生的一种新型电影形式。互动式电影是传统电影结合不断发展的数字游戏与数字媒体技术所产生的概念,它是相交于传统电影和电子游戏之间的新型领域,相较于电子游戏,互动式电影更加倾向于影像质感与整体叙事;相较于传统电影,游戏大师Chris Crawford认为:“电影并没有交互性。电影以精心构思好的方式进行叙事,而互动叙事则是随着情节的不断发展逐步建立起来的。”[ ChrisCrawford. 游戏大师Chris Crawford谈互动叙事[M]. 人民邮电出版社, 2015.] 互动式电影让观众拥有更多自由选择的权利,对于观众来说,互动式电影突破了以往的界限,观众不仅仅是观众,观众也成为了叙事的参与者与创造者。与往常观众被动接受内容的传统电影不同,互动式电影在不同的叙事情节点设置不同的个性化选项,观众可以根据自己的意愿从提供的交互选项中选择一个延续剧情的发展,观众的差异化选择会进入不同的故事情节,最后形成多元的结局。 值得注意的是,观众选择的互动选项不一定是关键情节的关键选择,也有许多日常的交互,比如由2018年由奈飞(Netflix)出品的互动式电影·《黑镜:潘达斯奈基》中主角父亲早上问主角吃什么早餐,主角决定听什么音乐,这些细节都可以由观众个性选择,不同的选择会影响后续的剧情走向与结局。充满创造性和个性化的电影叙事,这是互动式电影带给观众一种崭新的介入性强的观影体验。现代电子游戏注重游戏操作和游戏机制,电子游戏玩家获得成就感的方式是通过自身的游戏操作技术突破一道道游戏关卡最终实现胜利,而互动式电影更注重的是通过交互机制增强故事的表现力,它没有获得胜利的条件也没有方向和路线,无论观众如何选择都不会导致失败,最终的结局都是由观众根据自身的意愿选择产生,都是拥有闭环叙事的结局。 黑镜:潘达斯奈基 (二)互动式电影的发展现状 一部63分钟的捷克短片叫做《一个男人与他的房子》于1967年出现在蒙特利尔世博会上,惊动了当时整个电影界,这是世界上第一部互动式电影,当时电影的播放渠道主要集中在影院放映,配备按钮和互动设备的高额成本和技术限制让影院望而却步,很快互动式电影便陷入沉寂之中。 随着科学技术的发展,电子计算机,电子移动设备的大量使用以及互联网的不断发展,人类进入了数字媒介时代,随着3D图形技术的逐渐成熟,电子游戏电影化形成的交互式电影游戏崭露头角,在2010年,游戏制作人大卫·凯奇的作品《暴雨》 打破了互动式电影游戏的沉寂。 超凡双生 二、互动式电影跨媒介叙事方式的探究 (一)超叙事互动:网状及分支叙事模式 传统的电影叙事是导演对电影情节整体安排下的一次线性叙事,因为这种叙事单线输出于观众的。我们不难发现其实传统电影中的所有线索往往都涵盖在一个固定的框架中,犹如交错的道路,导演只设置一条路线所通往的地点,可能线索伴随着时间或是空间进行线性推动。“超叙事互动” 这种多选择,多路线的网状及分支叙事模式是互动式电影最大特点之一,观众的每一次选择都在原有的故事情节上建立了新平行时空,互动者在现实世界中思考和抉择,互动者的每一个选择都会影响虚拟世界的剧情走向,从而推进故事的发展。如互动式电影《夜班》。 (二)开放式叙事:被弱化的间离效应 传统电影中观众观看电影的元素包括:剧情,画面,演员表演,特效动画,都是由电影导演与编剧们创造的一个“囚笼”,观众们从电影开始到结束始终困在这个盒子里,无论电影如何演绎都无法突破这个“囚笼”,这是传统电影的“封闭叙事”属性。而互动式电影通过融合电子游戏的交互模式让观众拥有了打破“囚笼”的能力,观众拥有了书写文本的能力,打破了创作者们打造的囚笼,“开放叙事”由此产生。 上文提及的《一个男人与他的房子》需要观众选择红绿按钮来决定不同的电影剧情走向。观众们在观影时不同的选择给与了观众们新奇感,但完成所有的选择后影片的结局却又让观众们食而无味,观众以为自己选择的不同按钮能够给电影带来不同的结局走向,但是他们的选择毫无意义,无论如何选择那个按钮走那条剧情线,影片的结局最终也只有一个,直到最后观众们还是被导演玩弄于股掌之中。通过这个例子我们不难发现其实最早的互动式电影中还是无法脱离传统电影的线性叙事思路,即所有线索往往都涵盖在一个固定的框架中,犹如交错的道路。 导演预先设置一个所有道路所通往的终点,观众则通过自身思考,选择不同的道路。观众设想的每一条道路都可以使整个观影过程完整。但是观众面对的剧情选择往往情感上仍有缺失,因为导演可以设定路线但不能对每一位观众面对的结局进行优化,线性电影的结局往往是封闭的,唯一的,一部分原因归结于当时电影的时长和容量限制了这种需求的达成。如今随着互联网与新媒体技术的发展,互动式电影真正让观众参与了叙事,实现了影游在叙事上的融合。观众真正加入叙事,并可以按照自身的意志对叙事进行创作,此时不同方向的开展路线拥有了不同的结局,传统的封闭式结构由此发生了改变。如互动式电影《深海》共有2条主要故事线,15条支线,将近60个交互点,高达25个结局;《黑镜:潘达斯奈基》主副线约5个小时,30多个交互点相互排列组合最终会产生5个各不相同的方向的15个结局;《夜班》整部电影更拥有高达180个可以选择的交互按钮,7个不同的结局。面对以往观众对传统电影唯一结局的情感缺失,互动式电影给予了观众重新再来的契机,错误可以被改正,惋惜的瞬间可以被重现。 现阶段互动式电影的开放式叙事更多的是对叙事文本的不确定性进行填充,可进行的交互选择往往还是在一定的范围之内,但已经产生了观众对叙事文本的能动性参与,弱化了间离效果。间离化效果是由戏剧家布莱希特提出的戏剧表演艺术的代表性观点,在电影艺术中普遍得到应用,它主要表现为利用旁白、标语、歌唱等形式提醒观众当前的故事为虚构而不是真实,并通过客观视角或第三人称视角对叙事进行修改和更正。[ 贝・布莱希特. 布莱希特论戏剧[M]. 中国戏剧出版社, 1990.]由于互动式电影更注重沉浸感和主体的情感体验,在互动式电影中制作者往往就会弱化间离效果,通过设置交互界面增强观众的沉浸感体验。为了达到间离效果的目的,互动式电影往往采用分章节分段落的方式让观众逐渐了解电影的世界观和虚构的故事框架,最终完成对故事叙事的认知,形成开放式叙事。
1. “成就感”不可译为“achieve satisfaction”,推荐译为sense of achievement 2. “结局走向”不可译为“ending directions”,推荐译为“endings”或“outcomes” 3. “介入性”推荐译为“highly participatory”或“highly interactive”
4c360
学术论文
人文科学
193
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Representations, Warranties and Undertakings Party A hereby represents and warrants to Party B as follows:a) Party A is a company duly organized, validly existing and in good standing as a legal person under the laws of the PRC.b) Party A has full legal right, power and authority to execute and deliver this Contract and all of the contracts and documents referred to in this Contract to which Party A is a party and to observe and perform its obligations hereunder and thereunder. c)Party A has taken all appropriate and necessary corporate actions to authorize the execution and delivery of this Contract and all of the contracts and documents referred to in this Contract to which Party A is a party and to authorize the performance and observance of the terms and conditions hereof and thereof. d) Party A has obtained all consents, approvals and authorizations necessaryfor the valid execution and delivery of this Contract and all of the contractsand documents referred to in this Contract to which Party B is a party,provided, however, that this Contract is subject to the approval of the Examination and Approval Authority before the same may become effective. The Vendor represents, warrants and undertakes to and with the Purchaserthat each ofthe statements set out in Schedule A is now and shall at all timesbetween the date hereof and Completion (both dates inclusive) be true andaccurate. The Vendor acknowledges that the Purchaser has entered into thisAgreement in reliance upon the Warranties and has been induced by them toenter into this Agreement. Party A represents and warrants that there are no conditions at, on, underor related to, the real property constituting all or any portion of the Landwhich presently or potentially pose a hazard to human health or theenvironment, whether or not in compliance with law, and there has been nomanufacture, use, treatment, storage, transportation, or disposal of any hazardous or toxic substance, pollutant, or contaminant on the Land nor anyrelease of any hazardous or toxic substance, pollutant, or contaminant intoor upon or over the Land. Party A further represents and warrants that the Land is free and clear of anyand all claims, charges, easement, encumbrances, lease, covenants, securityinterest, liens, option, pledge, rights of others, or restrictions, whetherimposed by agreement, understanding, law, equity or otherwise. Party A warrants, agrees and undertakes to Party B that neither Party A(whether directly or indirectly) nor any Connected Person (whether individually or jointly) shall, during the term of this Agreement, and for two(2)years after the expiry or termination of this Agreement:a)undertake any business which is in direct competition with the business ofParty B without the written consent issued by the legal representative ofParty B; b) solicit or entice away any customer, employee, director or supplier ofParty B for whatever reason; and c)produce any products which is identical or similar in composition or appearance to any of the Products without the prior consent of Party B.For the purposes of this Article 4: "Connected Person" means any or all of the following persons: (i) the parent company or shareholders holding 50% or more of the equity of Party A; (ii)any corporation or business that is affiliated with Party A; or (iii) any officer employee, independent contractor, partner, joint venturer or agent of Party A or of any business affiliated with Party A; or (iv) any firm or corporation in which Party A or its parent company or affiliate company has an interest whether such interest is legally enforceable or not.5.The express warranties in this Agreement shall be in lieu of all other warranties, express or implied, including the implied warranties of interoperability, and fitness for a particular purpose.merchant ability, non-infringement, Responsibilities and Obligations Party A shall exercise, and ensure that all its servants, employees, agents or contractors exercise, all due diligence and care in the performance of the Services and each of the responsibilities so described. Specifically, andwithout limiting the generality of the foregoing, Party A shall bear allresponsibility for any losses or damages suffered by Party B as a result ofany mistakes, errors or omissions caused by Party A in connection with theprocessing and packing of the Products. The Company appreciates that whilst the Sponsor will use its reasonableendeavors to discharge its duties as sponsor to the Company and provide theCompany with advice and assistance as described above, it remains theprimary responsibility of the directors of the Company to ensure that theCompany will comply in full with and discharge its responsibili-ties underthe Growth Enterprise Market ("GEM")Listing Rules and other relevantlaws and regulations applicable to the Company. Save as describedhereunder, the Sponsor does not have any duty to monitor or otherwise toensure that the Company is in continuous compliance with the GEM ListingRules and other relevant laws and regulations applicable to the Company. 3.It is understood and agreed that each of the parties hereto is anindependent contractor and that neither party is, nor shall be considered tobe, an agent, distributor, fiduciary or representative of the other. Neitherparty shall act or represent itself, directly or by implication, as an agent ofthe other or in any manner assume or create any obligation on behalf of, orin the name of, the other. Responsibilities of Party A: In addition to its other obligations under this Contract, Party A shall, from time to time throughout the period of the Joint Venture term, be responsible. at no cost to the JVC (excepted as provided in Section X or as mayotherwise be agreed by the JVC in writing), for the following: a) to use best efforts to apply and take all actions necessary or appropriateon behalf of the JVC to obtain the necessary approvals, permits, certificatesand licenses for the establishment and operation ofthe JVC, and use the bestefforts to do, fulfill and perform any and all acts, conditions and thingsrequired to ensure the continuing validity and effectiveness of suchapprovals, permits, certificates and licenses;b)to assist the JVC in applying for, obtaining and maintaining tax(including without limitation income tax, tariffs, customs duties, excisetaxes, business tax, and value-added tax)reductions and exemptions andother investment incentives that may be available to the JVC; c)to assist the JVC in applying for, obtaining and maintaining specialbusiness opportunities, relationships or qualifications relating to or madeavailable by any gov-ernment departments or agencies, including withoutlimitation tax-advantaged import quotas, subsidies for purchase of local rawmaterials, and special toll processing or government supply contracts;d) to assist the JVC in obtaining water and power supplies, transportationand communications services and facilities, and use best efforts to obtain foithe JVC utility rates and quotas and other rights under the most favorableterms and conditions available to the JVC, and to liaise with the relevantauthorities effectively to achieve all ofthe above; e)to assist the directors and employees of the JVC and any employees ofParty B and the JVC's foreign contractors and consultants traveling to Chinain connection with the JVC's activities to obtain all necessary entry visaswork permits and residence permits; f)(f) to assist the JVC in preparing and presenting customs declarationshandling clearing procedures for machinery and equipment purchased orleased outside China by the JVC, applying for all available exemptions fromcustoms duty, value added tax, sales tax or other charges, and arranging fortransportation of the same within China;. g)to assist the JVC in obtaining loans and other credit facilities fromChinese banks or other financial institutions at the most favorable rates andterms available; and h) to handle other matters entrusted to it by the JVC. Severability and Waiver If any one or more of the provisions contained in this Agreement or anydocument executed in connection herewith shall .; be invalid, illegal, orunenforceable in any respect under any applicable law,(i) the validity.legality and enforce-j ability of the remaining provisions contained hereinor c therein shall not in any way be affected or impaired and shall remainin full force and effect; and (ii) the invalid, illegal or unenforceableprovision shall be replaced by a valid, legal and enforceable provision thatcomes closest to expressing the true intent of such invalid, illegal orunenforceable provision. If any of the provisions ofthis Agreement is held invalid or unenforceableand unless the invalidity or unenforceability thereof does substantialviolation to the underlying intent \- and sense of the remainder of thisAgreement, such invalidity or unenforceability shall not affect in any waythe validity and enforceability of any other provisions of this Agreementexcept those which the invalidated or unenforceable provisions comprise anintegral part of or are otherwise clearly inseparable from. That invalidity or unenforceability shall not affect any valid and enforceable application of the remaining provisions, and each such provision shall be deemed to beeffective, operative, made, or entered into in the manner and to the full extent permitted by law. Any provision of this Deed prohibited by or which is unlawful orunenforceable under any applicable law actually applied by any court ofcompetent jurisdiction shall, to the extent required by such law, be severecfrom this Deed and rendered ineffective so far as is possible withoutmodifying the remaining provisions of this Deed. Where, however, theprovisions of any such applicable law may be waived, they are herebywaived by the parties hereto to the full extent permitted by such law to theend that this Deed shall be valid, binding and enforceable in accordancewith its terms. Unless otherwise provided for herein, failure or delay on the part of anyparty to exercise any right, power or privilege under this Agreement shallnot operate as a waiver thereof, nor shall any single or partial exercise ofany right, power or privilege preclude further exercise thereof or exercise ofany other right, power or privilege. A waiver by one of the parties at anytime of a breach of any term or provision of this Agreement committed bythe other party shall not be construed as a waiver by such party of anysubsequent breach to be committed by the other party, nor shall it beconstrued as a waiver by such party of its rights under such provision or anyof its other rights under this Agreement. The failure of either party at any time or times to require performance ofany provision hereof shall in no manner affect its right at a later time toenforce the same. No waiver by either party of any condition or any breach of any of the terms, covenants or conditions contained in this Agreement shall be effective unless in writing, and no waiver in any one or more instances shall be construed as a further or continuing waiver of any other condition or any breach of any other terms, covenants or conditions.6. The Parties recognize and agree that their respective covenants andundertakings contained in this Agreement are of a special and unique natureand that a breach will result in irreparable injury for which there is noadequate remedy at law, and therefore the parties expressly agree that ifeither party shall at any time breach or in any way violate this Agreement.then Party A or Party B, as the case may be, shall be entitled to equitablerelief by way of injunction (in addition to, and not in substitution for, anyand all other relief to which such party may be entitled either at law or inequity) to restrain such breach and to compel compliance with theobligations undertaken. Each of the parties do hereby waive any proof that such breach will cause irreparable injury to such party or that there is no adequate remedy at law.
Representations, Warranties and Undertakings Party A hereby represents and warrants to Party B as follows:a) Party A is a company duly organized, validly existing and in good standing as a legal person under the laws of the PRC.b) Party A has full legal right, power and authority to execute and deliver this Contract and all of the contracts and documents referred to in this Contract to which Party A is a party and to observe and perform its obligations hereunder and thereunder. c)Party A has taken all appropriate and necessary corporate actions to authorize the execution and delivery of this Contract and all of the contracts and documents referred to in this Contract to which Party A is a party and to authorize the performance and observance of the terms and conditions hereof and thereof. d) Party A has obtained all consents, approvals and authorizations necessaryfor the valid execution and delivery of this Contract and all of the contractsand documents referred to in this Contract to which Party B is a party,provided, however, that this Contract is subject to the approval of the Examination and Approval Authority before the same may become effective. The Vendor represents, warrants and undertakes to and with the Purchaserthat each ofthe statements set out in Schedule A is now and shall at all timesbetween the date hereof and Completion (both dates inclusive) be true andaccurate. The Vendor acknowledges that the Purchaser has entered into thisAgreement in reliance upon the Warranties and has been induced by them toenter into this Agreement. Party A represents and warrants that there are no conditions at, on, underor related to, the real property constituting all or any portion of the Landwhich presently or potentially pose a hazard to human health or theenvironment, whether or not in compliance with law, and there has been nomanufacture, use, treatment, storage, transportation, or disposal of any hazardous or toxic substance, pollutant, or contaminant on the Land nor anyrelease of any hazardous or toxic substance, pollutant, or contaminant intoor upon or over the Land. Party A further represents and warrants that the Land is free and clear of anyand all claims, charges, easement, encumbrances, lease, covenants, securityinterest, liens, option, pledge, rights of others, or restrictions, whetherimposed by agreement, understanding, law, equity or otherwise. Party A warrants, agrees and undertakes to Party B that neither Party A(whether directly or indirectly) nor any Connected Person (whether individually or jointly) shall, during the term of this Agreement, and for two(2)years after the expiry or termination of this Agreement:a)undertake any business which is in direct competition with the business ofParty B without the written consent issued by the legal representative ofParty B; b) solicit or entice away any customer, employee, director or supplier ofParty B for whatever reason; and c)produce any products which is identical or similar in composition or appearance to any of the Products without the prior consent of Party B.For the purposes of this Article 4: "Connected Person" means any or all of the following persons: (i) the parent company or shareholders holding 50% or more of the equity of Party A; (ii)any corporation or business that is affiliated with Party A; or (iii) any officer employee, independent contractor, partner, joint venturer or agent of Party A or of any business affiliated with Party A; or (iv) any firm or corporation in which Party A or its parent company or affiliate company has an interest whether such interest is legally enforceable or not.5.The express warranties in this Agreement shall be in lieu of all other warranties, express or implied, including the implied warranties of interoperability, and fitness for a particular purpose.merchant ability, non-infringement, Responsibilities and Obligations Party A shall exercise, and ensure that all its servants, employees, agents or contractors exercise, all due diligence and care in the performance of the Services and each of the responsibilities so described. Specifically, andwithout limiting the generality of the foregoing, Party A shall bear allresponsibility for any losses or damages suffered by Party B as a result ofany mistakes, errors or omissions caused by Party A in connection with theprocessing and packing of the Products. The Company appreciates that whilst the Sponsor will use its reasonableendeavors to discharge its duties as sponsor to the Company and provide theCompany with advice and assistance as described above, it remains theprimary responsibility of the directors of the Company to ensure that theCompany will comply in full with and discharge its responsibili-ties underthe Growth Enterprise Market ("GEM")Listing Rules and other relevantlaws and regulations applicable to the Company. Save as describedhereunder, the Sponsor does not have any duty to monitor or otherwise toensure that the Company is in continuous compliance with the GEM ListingRules and other relevant laws and regulations applicable to the Company. 3.It is understood and agreed that each of the parties hereto is anindependent contractor and that neither party is, nor shall be considered tobe, an agent, distributor, fiduciary or representative of the other. Neitherparty shall act or represent itself, directly or by implication, as an agent ofthe other or in any manner assume or create any obligation on behalf of, orin the name of, the other. Responsibilities of Party A: In addition to its other obligations under this Contract, Party A shall, from time to time throughout the period of the Joint Venture term, be responsible. at no cost to the JVC (excepted as provided in Section X or as mayotherwise be agreed by the JVC in writing), for the following: a) to use best efforts to apply and take all actions necessary or appropriateon behalf of the JVC to obtain the necessary approvals, permits, certificatesand licenses for the establishment and operation ofthe JVC, and use the bestefforts to do, fulfill and perform any and all acts, conditions and thingsrequired to ensure the continuing validity and effectiveness of suchapprovals, permits, certificates and licenses;b)to assist the JVC in applying for, obtaining and maintaining tax(including without limitation income tax, tariffs, customs duties, excisetaxes, business tax, and value-added tax)reductions and exemptions andother investment incentives that may be available to the JVC; c)to assist the JVC in applying for, obtaining and maintaining specialbusiness opportunities, relationships or qualifications relating to or madeavailable by any gov-ernment departments or agencies, including withoutlimitation tax-advantaged import quotas, subsidies for purchase of local rawmaterials, and special toll processing or government supply contracts;d) to assist the JVC in obtaining water and power supplies, transportationand communications services and facilities, and use best efforts to obtain foithe JVC utility rates and quotas and other rights under the most favorableterms and conditions available to the JVC, and to liaise with the relevantauthorities effectively to achieve all ofthe above; e)to assist the directors and employees of the JVC and any employees ofParty B and the JVC's foreign contractors and consultants traveling to Chinain connection with the JVC's activities to obtain all necessary entry visaswork permits and residence permits; f)(f) to assist the JVC in preparing and presenting customs declarationshandling clearing procedures for machinery and equipment purchased orleased outside China by the JVC, applying for all available exemptions fromcustoms duty, value added tax, sales tax or other charges, and arranging fortransportation of the same within China;. g)to assist the JVC in obtaining loans and other credit facilities fromChinese banks or other financial institutions at the most favorable rates andterms available; and h) to handle other matters entrusted to it by the JVC. Severability and Waiver If any one or more of the provisions contained in this Agreement or anydocument executed in connection herewith shall .; be invalid, illegal, orunenforceable in any respect under any applicable law,(i) the validity.legality and enforce-j ability of the remaining provisions contained hereinor c therein shall not in any way be affected or impaired and shall remainin full force and effect; and (ii) the invalid, illegal or unenforceableprovision shall be replaced by a valid, legal and enforceable provision thatcomes closest to expressing the true intent of such invalid, illegal orunenforceable provision. If any of the provisions ofthis Agreement is held invalid or unenforceableand unless the invalidity or unenforceability thereof does substantialviolation to the underlying intent \- and sense of the remainder of thisAgreement, such invalidity or unenforceability shall not affect in any waythe validity and enforceability of any other provisions of this Agreementexcept those which the invalidated or unenforceable provisions comprise anintegral part of or are otherwise clearly inseparable from. That invalidity or unenforceability shall not affect any valid and enforceable application of the remaining provisions, and each such provision shall be deemed to beeffective, operative, made, or entered into in the manner and to the full extent permitted by law. Any provision of this Deed prohibited by or which is unlawful orunenforceable under any applicable law actually applied by any court ofcompetent jurisdiction shall, to the extent required by such law, be severecfrom this Deed and rendered ineffective so far as is possible withoutmodifying the remaining provisions of this Deed. Where, however, theprovisions of any such applicable law may be waived, they are herebywaived by the parties hereto to the full extent permitted by such law to theend that this Deed shall be valid, binding and enforceable in accordancewith its terms. Unless otherwise provided for herein, failure or delay on the part of anyparty to exercise any right, power or privilege under this Agreement shallnot operate as a waiver thereof, nor shall any single or partial exercise ofany right, power or privilege preclude further exercise thereof or exercise ofany other right, power or privilege. A waiver by one of the parties at anytime of a breach of any term or provision of this Agreement committed bythe other party shall not be construed as a waiver by such party of anysubsequent breach to be committed by the other party, nor shall it beconstrued as a waiver by such party of its rights under such provision or anyof its other rights under this Agreement. The failure of either party at any time or times to require performance ofany provision hereof shall in no manner affect its right at a later time toenforce the same. No waiver by either party of any condition or any breach of any of the terms, covenants or conditions contained in this Agreement shall be effective unless in writing, and no waiver in any one or more instances shall be construed as a further or continuing waiver of any other condition or any breach of any other terms, covenants or conditions.6. The Parties recognize and agree that their respective covenants andundertakings contained in this Agreement are of a special and unique natureand that a breach will result in irreparable injury for which there is noadequate remedy at law, and therefore the parties expressly agree that ifeither party shall at any time breach or in any way violate this Agreement.then Party A or Party B, as the case may be, shall be entitled to equitablerelief by way of injunction (in addition to, and not in substitution for, anyand all other relief to which such party may be entitled either at law or inequity) to restrain such breach and to compel compliance with theobligations undertaken. Each of the parties do hereby waive any proof that such breach will cause irreparable injury to such party or that there is no adequate remedy at law.
考点1:“duly organized, validly existing and in good standing”应整体视为法律惯用表达,推荐译为“依法设立、有效存续且资格完备”。 考点2:“observe and perform”常见于合同义务条款中,推荐译为“遵守并履行”,前者偏规范性、后者偏操作性。 考点3:“solicit or entice away”意群应理解为“招揽或诱使……离开”,勿拆译为“请求”或“诱惑”。 考点4:“hazardous or toxic substance, pollutant, or contaminant”为环境法术语,“contaminant"应译为“致污物”。 考点5:“equitable relief by way of injunction”中“injunction”应译为“禁令”。 考点6: “servant”在此语境中推荐译为“服务人员”,不可译为“仆人” 考点7: “understanding”在此语境中应译为“谅解”,属于法律文件中的固定翻译,不可译为“理解” 考点8: “charges”作为法律术语时,常指财产上的担保设定,这里应译为“担保”,不可译为“费用”
4e21b
垂类场景
法律
166
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: In recent years, with the rapid development of technologies such as artificial intelligence (AI), the Internet of Things (IoT), and 5G communication, a new wave of information technology revolution has emerged. AI technology, serving as the "central brain" that powers smart manufacturing, smart cities, and the metaverse, plays a pivotal role in the digital era. It is widely acknowledged that algorithms, data, and computational power (chips) are the three core elements driving AI development, with chips forming the foundation for the practical implementation of AI. As research on large models such as ChatGPT and GPT-4 continues to progress, model structures are becoming increasingly complex, and the demand for data and computational resources is growing exponentially. Simultaneously, the gradual diminishing of Moore’s Law has slowed the evolution of chip manufacturing processes. The disparity between algorithm development and hardware advancement poses a significant challenge: how to efficiently process massive datasets using complex algorithms has become a pressing issue in the AI domain. The root cause of this challenge lies in the limitations of chips, which are currently constrained by the "von Neumann architecture bottleneck." In von Neumann architecture, data storage and processing are separated, and data is transferred between memory and processors through buses. Frequent data transfers incur substantial time and energy costs, leading to what is termed the "memory wall" and the "power wall." Moreover, as device sizes approach physical limits, the traditional approach of improving chip performance solely through process advancements faces increasing obstacles, resulting in the so-called "process wall." To address these challenges, academia and industry worldwide have conducted extensive research from multiple dimensions, including architecture, process, and integration, to explore new-generation chip technologies for the post-Moore era. For instance, dataflow architecture chips enable streaming computation, achieving significantly higher throughput than von Neumann architecture for large-scale AI data processing; reconfigurable chips provide high flexibility and energy-efficient computation through software-defined hardware circuit structures; wafer-scale chips expand chip area using advanced process technologies to enhance computational power; and 3D chips stack multiple chips vertically using 3D integration packaging technology to achieve high bandwidth and computational power. Among these, in-memory computing chips stand out by integrating storage and computation through collaborative innovation across devices, architectures, circuits, and processes, fundamentally overcoming the von Neumann architecture bottleneck. In-memory computing chips have architectural features that significantly reduce data transfer overhead, addressing the "memory wall" and "power wall." Additionally, their highly parallel computing capabilities enable performance comparable to that of advanced processes, even with relatively mature process nodes, thereby alleviating the pressure of process miniaturization. Furthermore, in-memory computing technology seamlessly integrates with other advanced technologies, such as reconfigurable chips, wafer-scale chips, and 3D integration, making it one of the most promising directions in post-Moore era chip technology. The concept of in-memory computing can be traced back to the 1970s when Kautz from the Stanford Research Institute first proposed the idea. The core principle is to perform computation directly within memory, reducing the overhead of data transfer between memory and processors. Subsequent research has focused on chip circuit structures, computational architectures, and system applications. However, due to the complexity of circuit design and the difficulty of implementation, most of the early research effectively realized "near-memory computing." The main distinction between near-memory and in-memory computing is that the former still requires data to be read out of memory for computation, with the results then stored back into memory. Current industry-standard solutions for near-memory computing include techniques such as 3D packaging and high-bandwidth memory (HBM) to shorten the distance between memory and processors and increase data bandwidth. These technologies are relatively mature and have been scaled into mass production. Leading semiconductor companies such as AMD, Intel, Samsung, and SK Hynix have introduced near-memory computing chips based on HBM and 2.5D/3D packaging technologies. For example, Samsung’s latest HBM3 Icebolt technology employs a near-memory computing architecture, utilizing 12 layers of 10nm-class dynamic random-access memory (DRAM) to achieve processing speeds of up to 6.4 Gbps and bandwidths of up to 819 GB/s. However, near-memory computing still adheres to the von Neumann architecture, which fundamentally separates storage and computation. While it reduces data transfer overhead to some extent, it cannot entirely eliminate the von Neumann bottleneck. In recent years, in-memory computing architectures that integrate storage and computation into a unified framework have become a hotspot in academia and industry. In-memory computing chips can be broadly categorized into analog and digital paradigms: 1. Analog in-memory computing: Signals within or near the storage array are processed as analog signals. Operations such as matrix multiplications leverage the intrinsic physical behavior of the storage array, often relying on digital-to-analog and analog-to-digital converters for signal conversion. Analog computation typically utilizes Ohm’s law and Kirchhoff’s law to achieve high energy efficiency but may suffer from lower precision and reliability. 2. Digital in-memory computing: Signals within or near the storage array are processed as digital signals, typically using techniques like adder trees and shifters. Digital in-memory computing achieves arbitrary precision but has relatively lower energy efficiency and higher cost per unit area. Research on in-memory computing chips spans various storage media. Chips based on traditional media like SRAM and NOR Flash are relatively mature and have entered small-scale production. Meanwhile, chips leveraging emerging non-volatile storage media, such as magnetic random-access memory (MRAM), resistive random-access memory (RRAM), phase-change random-access memory (PCRAM), and ferroelectric field-effect transistors (FeFET), are transitioning from foundational research to industrialization. Research institutions and companies from the United States, China, and South Korea have made significant advances in this field, with notable contributions from IBM, Stanford University, Princeton University, SK Hynix, Samsung, Tsinghua University, Peking University, and the Chinese Academy of Sciences Institute of Microelectronics. Despite these advancements, effective chip design relies heavily on EDA (Electronic Design Automation) tools, which remain in the early stages for in-memory computing. Current academic efforts have produced simple single-point tools for simulation modeling and automated synthesis, but no comprehensive EDA tool platform akin to those for digital circuit design has been established, making it challenging to support large-scale automated design of in-memory computing chips. Current research on in-memory computing chips is primarily focused on single-point technologies, with many technical challenges remaining in areas such as devices, circuits, architectures, EDA tools, and system applications. The future development of in-memory computing chips is expected to focus on the following four trends: 1. Shifting Toward the Perception End for Ultra-Low Power Consumption: Targeting edge markets like wearable devices and IoT, ultra-low power and low-cost solutions will be developed. Current perception chips rely on analog-to-digital converters to process data, which is slow and energy-intensive. By integrating in-memory computing chips with perception modules and developing efficient mixed-signal processing methods, it will be possible to achieve a fully integrated perception-storage-computation solution, drastically reducing overhead and achieving ultra-low power consumption. 2. Shifting Toward the Edge/Cloud for Extreme Computational Power: For applications like edge/cloud servers, data centers, and autonomous driving, in-memory computing chips can offer massive parallel computing capabilities. Current edge/cloud processors, typically based on GPUs, face significant data communication overheads, limiting their actual computational efficiency. In-memory computing chips, optimized for large-scale matrix operations such as those in Transformer-based models, can greatly reduce bandwidth requirements and enhance computational power. 3. Co-Design of Heterogeneous Architectures and Integration: Combining different computational architectures and hardware units can maximize their strengths while compensating for their weaknesses. For example, integrating analog and digital in-memory computing architectures can strike a balance between precision, energy efficiency, area, and cost. Advanced packaging technologies like 2.5D/3D integration and chiplets can be leveraged to integrate CPUs, GPUs, and in-memory computing chips fabricated at different process nodes, fully realizing the advantages of heterogeneous architectures. 4. Development of EDA and Application Toolchains: As in-memory computing chips move from proof-of-concept to large-scale production, the development of automated EDA tools and application toolchains becomes critical. These tools can simplify chip design, reduce development costs, and accelerate deployment. A robust ecosystem of open-source and standardized platforms will drive mass production and widespread adoption of in-memory computing chips. To date, in-memory computing chip technology has garnered significant attention from academia and industry, achieving a series of breakthroughs. However, before large-scale industrialization can be realized, several critical scientific issues need to be addressed in areas such as low-power mixed-signal circuits, high-density in-memory computing media, automated EDA tools, and methods/standards/processes for heterogeneous integration. At the terminal end, the core challenge lies in the ultra-low-power processing of multi-modal information. Existing solutions for perception-storage-computing chips primarily focus on single-modality signal processing (e.g., auditory, visual, or tactile), while intelligent systems often require comprehensive analysis of multi-modal data to perceive and interpret their environment. Designing a multi-modal perception-storage-computing architecture and its associated low-power mixed-signal circuits is a critical scientific problem in achieving "ultra-low power" integration. This research spans multiple disciplines, including devices, processes, chips, algorithms, and applications, necessitating cross-layer collaborative optimization. Key questions include improving existing devices or fabricating new ones, optimizing processing technology for reliability and precision, designing efficient multi-modal mixed-signal circuits, creating algorithms tailored to hardware, and establishing user-friendly application ecosystems. At the edge/cloud level, the primary challenges are computational power, precision, and bandwidth. Leveraging the massive parallel computing capabilities of in-memory computing chips becomes particularly critical for large-scale model applications, which impose stringent demands on scale, precision, and computational power. For instance, large-scale models require single arrays exceeding 10k × 10k, total memory capacities reaching gigabit levels, and single-chip computational performance (8-bit fixed-point numbers) of at least 100 TOPS (tera operations per second). Furthermore, increasing integration scales bring challenges like error propagation, non-uniformity, and reliability issues. Overcoming these requires breakthroughs in fundamental principles, device technologies, and theoretical models for defect and error propagation.
In recent years, with the rapid development of technologies such as artificial intelligence (AI), the Internet of Things (IoT), and 5G communication, a new wave of information technology revolution has emerged. AI technology, serving as the "central brain" that powers smart manufacturing, smart cities, and the metaverse, plays a pivotal role in the digital era. It is widely acknowledged that algorithms, data, and computational power (chips) are the three core elements driving AI development, with chips forming the foundation for the practical implementation of AI. As research on large models such as ChatGPT and GPT-4 continues to progress, model structures are becoming increasingly complex, and the demand for data and computational resources is growing exponentially. Simultaneously, the gradual diminishing of Moore’s Law has slowed the evolution of chip manufacturing processes. The disparity between algorithm development and hardware advancement poses a significant challenge: how to efficiently process massive datasets using complex algorithms has become a pressing issue in the AI domain. The root cause of this challenge lies in the limitations of chips, which are currently constrained by the "von Neumann architecture bottleneck." In von Neumann architecture, data storage and processing are separated, and data is transferred between memory and processors through buses. Frequent data transfers incur substantial time and energy costs, leading to what is termed the "memory wall" and the "power wall." Moreover, as device sizes approach physical limits, the traditional approach of improving chip performance solely through process advancements faces increasing obstacles, resulting in the so-called "process wall." To address these challenges, academia and industry worldwide have conducted extensive research from multiple dimensions, including architecture, process, and integration, to explore new-generation chip technologies for the post-Moore era. For instance, dataflow architecture chips enable streaming computation, achieving significantly higher throughput than von Neumann architecture for large-scale AI data processing; reconfigurable chips provide high flexibility and energy-efficient computation through software-defined hardware circuit structures; wafer-scale chips expand chip area using advanced process technologies to enhance computational power; and 3D chips stack multiple chips vertically using 3D integration packaging technology to achieve high bandwidth and computational power. Among these, in-memory computing chips stand out by integrating storage and computation through collaborative innovation across devices, architectures, circuits, and processes, fundamentally overcoming the von Neumann architecture bottleneck. In-memory computing chips have architectural features that significantly reduce data transfer overhead, addressing the "memory wall" and "power wall." Additionally, their highly parallel computing capabilities enable performance comparable to that of advanced processes, even with relatively mature process nodes, thereby alleviating the pressure of process miniaturization. Furthermore, in-memory computing technology seamlessly integrates with other advanced technologies, such as reconfigurable chips, wafer-scale chips, and 3D integration, making it one of the most promising directions in post-Moore era chip technology. The concept of in-memory computing can be traced back to the 1970s when Kautz from the Stanford Research Institute first proposed the idea. The core principle is to perform computation directly within memory, reducing the overhead of data transfer between memory and processors. Subsequent research has focused on chip circuit structures, computational architectures, and system applications. However, due to the complexity of circuit design and the difficulty of implementation, most of the early research effectively realized "near-memory computing." The main distinction between near-memory and in-memory computing is that the former still requires data to be read out of memory for computation, with the results then stored back into memory. Current industry-standard solutions for near-memory computing include techniques such as 3D packaging and high-bandwidth memory (HBM) to shorten the distance between memory and processors and increase data bandwidth. These technologies are relatively mature and have been scaled into mass production. Leading semiconductor companies such as AMD, Intel, Samsung, and SK Hynix have introduced near-memory computing chips based on HBM and 2.5D/3D packaging technologies. For example, Samsung’s latest HBM3 Icebolt technology employs a near-memory computing architecture, utilizing 12 layers of 10nm-class dynamic random-access memory (DRAM) to achieve processing speeds of up to 6.4 Gbps and bandwidths of up to 819 GB/s. However, near-memory computing still adheres to the von Neumann architecture, which fundamentally separates storage and computation. While it reduces data transfer overhead to some extent, it cannot entirely eliminate the von Neumann bottleneck. In recent years, in-memory computing architectures that integrate storage and computation into a unified framework have become a hotspot in academia and industry. In-memory computing chips can be broadly categorized into analog and digital paradigms: 1. Analog in-memory computing: Signals within or near the storage array are processed as analog signals. Operations such as matrix multiplications leverage the intrinsic physical behavior of the storage array, often relying on digital-to-analog and analog-to-digital converters for signal conversion. Analog computation typically utilizes Ohm’s law and Kirchhoff’s law to achieve high energy efficiency but may suffer from lower precision and reliability. 2. Digital in-memory computing: Signals within or near the storage array are processed as digital signals, typically using techniques like adder trees and shifters. Digital in-memory computing achieves arbitrary precision but has relatively lower energy efficiency and higher cost per unit area. Research on in-memory computing chips spans various storage media. Chips based on traditional media like SRAM and NOR Flash are relatively mature and have entered small-scale production. Meanwhile, chips leveraging emerging non-volatile storage media, such as magnetic random-access memory (MRAM), resistive random-access memory (RRAM), phase-change random-access memory (PCRAM), and ferroelectric field-effect transistors (FeFET), are transitioning from foundational research to industrialization. Research institutions and companies from the United States, China, and South Korea have made significant advances in this field, with notable contributions from IBM, Stanford University, Princeton University, SK Hynix, Samsung, Tsinghua University, Peking University, and the Chinese Academy of Sciences Institute of Microelectronics. Despite these advancements, effective chip design relies heavily on EDA (Electronic Design Automation) tools, which remain in the early stages for in-memory computing. Current academic efforts have produced simple single-point tools for simulation modeling and automated synthesis, but no comprehensive EDA tool platform akin to those for digital circuit design has been established, making it challenging to support large-scale automated design of in-memory computing chips. Current research on in-memory computing chips is primarily focused on single-point technologies, with many technical challenges remaining in areas such as devices, circuits, architectures, EDA tools, and system applications. The future development of in-memory computing chips is expected to focus on the following four trends: 1. Shifting Toward the Perception End for Ultra-Low Power Consumption: Targeting edge markets like wearable devices and IoT, ultra-low power and low-cost solutions will be developed. Current perception chips rely on analog-to-digital converters to process data, which is slow and energy-intensive. By integrating in-memory computing chips with perception modules and developing efficient mixed-signal processing methods, it will be possible to achieve a fully integrated perception-storage-computation solution, drastically reducing overhead and achieving ultra-low power consumption. 2. Shifting Toward the Edge/Cloud for Extreme Computational Power: For applications like edge/cloud servers, data centers, and autonomous driving, in-memory computing chips can offer massive parallel computing capabilities. Current edge/cloud processors, typically based on GPUs, face significant data communication overheads, limiting their actual computational efficiency. In-memory computing chips, optimized for large-scale matrix operations such as those in Transformer-based models, can greatly reduce bandwidth requirements and enhance computational power. 3. Co-Design of Heterogeneous Architectures and Integration: Combining different computational architectures and hardware units can maximize their strengths while compensating for their weaknesses. For example, integrating analog and digital in-memory computing architectures can strike a balance between precision, energy efficiency, area, and cost. Advanced packaging technologies like 2.5D/3D integration and chiplets can be leveraged to integrate CPUs, GPUs, and in-memory computing chips fabricated at different process nodes, fully realizing the advantages of heterogeneous architectures. 4. Development of EDA and Application Toolchains: As in-memory computing chips move from proof-of-concept to large-scale production, the development of automated EDA tools and application toolchains becomes critical. These tools can simplify chip design, reduce development costs, and accelerate deployment. A robust ecosystem of open-source and standardized platforms will drive mass production and widespread adoption of in-memory computing chips. To date, in-memory computing chip technology has garnered significant attention from academia and industry, achieving a series of breakthroughs. However, before large-scale industrialization can be realized, several critical scientific issues need to be addressed in areas such as low-power mixed-signal circuits, high-density in-memory computing media, automated EDA tools, and methods/standards/processes for heterogeneous integration. At the terminal end, the core challenge lies in the ultra-low-power processing of multi-modal information. Existing solutions for perception-storage-computing chips primarily focus on single-modality signal processing (e.g., auditory, visual, or tactile), while intelligent systems often require comprehensive analysis of multi-modal data to perceive and interpret their environment. Designing a multi-modal perception-storage-computing architecture and its associated low-power mixed-signal circuits is a critical scientific problem in achieving "ultra-low power" integration. This research spans multiple disciplines, including devices, processes, chips, algorithms, and applications, necessitating cross-layer collaborative optimization. Key questions include improving existing devices or fabricating new ones, optimizing processing technology for reliability and precision, designing efficient multi-modal mixed-signal circuits, creating algorithms tailored to hardware, and establishing user-friendly application ecosystems. At the edge/cloud level, the primary challenges are computational power, precision, and bandwidth. Leveraging the massive parallel computing capabilities of in-memory computing chips becomes particularly critical for large-scale model applications, which impose stringent demands on scale, precision, and computational power. For instance, large-scale models require single arrays exceeding 10k × 10k, total memory capacities reaching gigabit levels, and single-chip computational performance (8-bit fixed-point numbers) of at least 100 TOPS (tera operations per second). Furthermore, increasing integration scales bring challenges like error propagation, non-uniformity, and reliability issues. Overcoming these requires breakthroughs in fundamental principles, device technologies, and theoretical models for defect and error propagation.
考点1:“memory wall” 必须译为 “存储墙”。 考点2:“process wall” 必须译为 “制程墙”。 考点3:“High‑Bandwidth Memory (HBM)” 必须译为 “高带宽内存(HBM)”。 考点4:“analog in‑memory computing” 必须译为 “模拟存内计算”。 考点5:“digital in‑memory computing” 必须译为 “数字存内计算”。 考点6:“SRAM / NOR Flash” 必须译为 “静态随机存取存储器 / NOR型闪存”。 考点7:“MRAM / RRAM / PCRAM / FeFET” 必须译为 “磁阻存储器 / 阻变存储器 / 相变存储器 / 铁电场效应晶体管”。 考点8:“chiplets / 2.5D/3D integration” 必须译为 “芯粒 / 2.5D/3D集成”。 考点9:“dynamic random‑access memory (DRAM)” 必须译为 “动态随机存取存储器(DRAM)”。 考点10:“TOPS (8-bit fixed-point)”推荐译为“万亿次每秒(8位定点,TOPS)”
4f73b
学术论文
应用学科
150
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Cave 17 at the Dunhuang cave complex is a cache of religious texts and social records sealed at the beginning of the eleventh century and discovered in 1900. Legal documents from this cave show that ordained Buddhist monks and nuns in Dunhuang actively engaged in legal practices as plaintiffs, defendants, witnesses, mediators, and representatives in the trial of various disputes. Where did these monks and nuns get their legal education? And what roles did they play in circulating legal knowledge in Dunhuang? Addressing these questions is crucial for my larger project on the interaction between religion and law in China as they can help paint a fuller picture of how Buddhism and the state interacted in the legal arena in pre-modern China. Elsewhere, I have discussed how the jurisdictional boundary between the state and the Buddhist monastic community shifted in central China and analyzed cases in which individual monks and nuns litigated in the lay courts in Dunhuang. This article aims to understand the legal landscape on the frontier in Dunhuang, with a particular focus on ordained Buddhists’ access to legal texts and the roles they played in producing and preserving them. My primary sources are colophons on the Chinese manuscripts of Buddhist monastic law texts (Vinaya) discovered in the Dunhuang cave 17. My examination of the said colophons will reveal that ordained Buddhist monks and nuns played various roles in the circulation of legal knowledge in Dunhuang. They were patrons, scribes, and owners of various Buddhist monastic law texts circulated in Dunhuang. In cave 17, we can find Buddhist monastic law texts of all the major surviving traditions. The majority of them are law texts transmitted from the Dharmaguptaka tradition, but those of the Mahāsāṃghika, Sarvāstivāda, Mahīśāsaka, and Mūlasarvāstivāda traditions have also appeared in the same cave. These monastic law texts cover Buddhist liturgical and disciplinary guidelines ranging from the five precepts for the lay Buddhist householder, to the eight temporary fasting precepts, the 10 precepts for ordained novices, the six additional precepts for probationary Buddhist nuns, and the full set of disciplinary rules for fully ordained Buddhist monks and nuns. This examination, however, is not comprehensive. Ideally, we need to combine this examination with a study on the Buddhist monastic law texts written in Tibetan and other languages found in cave 17. Since Shayne Clarke and his team at McMaster University are currently working on a project surveying the Tibetan Vinaya texts from the Dunhuang manuscript collection, I did not discuss the Tibetan materials in the present article. It is also important to note that while my discussion below focuses on Buddhist monastic law texts, those were not the only type of legal texts circulated in Dunhuang. In this cave, we also find fragments of the primary legal texts promulgated by the Tang court. These include code, statutes, regulations, and ordinances. Several judicial judgments from the Tang, one of which I will discuss below, were also found in Dunhuang. Moreover, we have also found damaged copies of the section on The Penal Code from The Book of the Han. Before discussing how ordained Buddhists engaged in the copying, collecting, and preserving of Buddhist monastic law texts in Dunhuang, it is necessary to provide an overview of the workflow of text copying in Dunhuang. Scribes in Dunhuang copied Buddhist legal texts over several centuries. The earliest Buddhist monastic law text found in cave 17 is the Sarvāstivāda Prātimokṣa for Ordained Buddhist Monks . The scribe, a Buddhist monk Deyou, stated in the colophon that he copied this text in the southern part of Dunhuang during the summer rainy season retreat in 406, one year after he received his full ordination in 405. Copying of Buddhist monastic law texts continued in Dunhuang without interruption toward the end of the tenth century. Even after Dunhuang fell under Tibetan control, Buddhist text copying continued from 786 to 848. The Tibetan scholar Zhang Yanqing’s extensive study on the copying of Tibetan Buddhist texts in Dunhuang helps unveil important aspects of the workflow of text copying in this area. At the beginning of the Tibetan occupation of Dunhuang, text-copying focused on Chinese Buddhist texts. With support from the Tibetan king Khri gtsug lde brtsan, massive copying of Tibetan Buddhist texts began in 822 and continued in Shazhou until 838. Teams stationed at a particular scriptorium (gur) with multiple scribes and proofreaders copied large texts. Before starting the copying, they would secure a mother copy (dpe’) and proofread it multiple times. Each scribe would receive a mother copy of their targeted text and paper, and an administrator would record the distribution of paper and mother copies. A loan record on Pelliot chinois 3010v illustrates this practice. This manuscript states that on the seventeenth day in the fifth month in a mouse year , Sanjie monastery, one of the local monasteries in Dunhuang, borrowed six bundles of the Chinese translation of the Dharmaguptaka Vinaya and their brocade bundle covers for copying . When a scribe completed a copy, a proofreader would check the replica against the original. It was normal to proofread the copied Tibetan Buddhist texts at least three or four times, or 10 times in rare cases. In contrast to proofreading the Buddhist text commissioned by the Tibetan king anywhere between three times and 10 times, proofreaders normally proofread the Chinese copies of Buddhist monastic law texts found in Dunhuang once, twice, or none. For instance, Jingtu monastery owned a copy of the Commentary on the Dharmaguptaka Prātimokṣa. Its colophon informs us that an unspecified proofreader proofread it once at the Mogao cave on the first day in the third month in a horse year. Copies of the Buddhist monastic law texts found in cave 17 were produced in various locations. Some of them were locally copied in Dunhuang. On the nineteenth day in the eleventh month in a dog year, an unspecified scribe copied A Short Collection of Buddhist Monastic Law Texts at Bao’en monastery in Dunhuang. On the twenty-eighth day in the third month in a mouse year, another unspecified scribe copied the third scroll of the Commentary on the Dharmaguptaka Prātimokṣa at Yongshou monastery in Shazhou. This copy was proofread and punctuated in red ink. On the twentieth day in the tenth month in the seventh year of a certain era, a Buddhist monk, Riding , copied The Dharmaguptaka Prātimokṣa for Ordained Buddhist Monks at Dayun monastery. In the twelfth month in a mouse year, a monk from Jinguangming monastery, who was also a Vinaya master, copied the Commentary on the Dharmaguptaka Prātimokṣa for circulation. On the twentieth day in the fourth month in the second year during the Qianyuan era (759), monk Jingshen from Longxing monastery copied the Dharmaguptaka Prātimokṣa for Ordained Buddhist Monks. Another Longxing monastery monk, Zhizhao, copied the Dharmaguptaka Prātimokṣa for Ordained Buddhist Nuns. From the eighth day in the fifth month to the third day in the sixth month in a horse year, a Buddhist monk, Liji, from Jinguangming monastery copied one scroll of the Tang Buddhist monk Daoxuans (596–667) Randomly Selected Ecclesiastical Proceedings: Abridgements and Emendations to the Dharmaguptaka Vinaya for his monastery’s elder monk, Jinyao, at the beginning of that year’s rainy season retreat. Another local site where multiple Buddhist legal texts were copied is the Dongshan solitary monastic dwelling, which was most likely one of many such dwellings built near the Dunhuang cave complex. At least three surviving Buddhist law texts were copied at this site in a tiger year. On the sixteenth day in the sixth month in a tiger year, an unspecified individual completed proofreading the Notes on Conduct: Abridgements and Emendations to the Dharmaguptaka Vinaya at the Dongshan solitary monastic dwelling. One month later, on the fifteenth day in the seventh month of a tiger year, an unspecified scribe finished copying a single-scroll text, A Concise Copy of the Vinaya at Dongshan, identifiable with the Dongshan solitary monastic dwelling mentioned above. The owner of this copied text was a Buddhist nun, Shengzang. A few months later, on the twentieth day in the tenth month in a tiger year, an unspecified scribe continued copying and proofreading the fourth scroll of the Commentary on the Dharmaguptaka Prātimokṣa. A bit further away from Dunhuang, the Xiuduo monastery in Ganzhou Prefecture in todays Gansu province was also an important site for the circulation of legal and other Buddhist knowledge in Dunhuang. A Buddhist monk, Huizhen, copied The Sūtra on the Twenty-four Precepts for the Faithful Bodhisattvas Told by the Buddha at this Xiuduo monastery. The date was unspecified. This monastery was also where the Tibetan Buddhist monk, Chos grub (Ch. Facheng), taught and translated texts. Turfan, located about 800 kilometers to the Northwest of Dunhuang, appeared to be another productive center for copying Buddhist law texts. On the fourth day in the seventh month in the second year during the Guangde era, a Buddhist monk, Yilin, copied Daoxuans Notes on Conduct: Abridgements and Emendations to the Dharmaguptaka Vinaya at a place specified as Peijiata in the Nanping city in the Xizhou Prefecture, in today’s Turfan. At the end of the third and last scroll of this text, the colophons mention the scribe’s name, where and when he copied it. At the end of the middle scroll of this text, an unspecified proofreader wrote in red ink that he proofread the copied text in the early half of the sixth month in a snake year at the garden of a central Asian mound, whose exact location remains unknown. Because the nearest snake year after 764 was 765, this text was likely proofread in 765. Some of the Buddhist monastic law texts traveled to Dunhuang from central China. In the Dunhuang cave 17, we found at least two Buddhist monastic law texts previously owned by ordained Buddhists in central China. One is The Manual for Ordained Buddhists to Receive the Bodhisattva Precepts, which was copied under imperial order in the fifth month in the eighteenth year during the Tianjian era (519) during the Liang Dynasty. A layperson, Dai Mengtong, was the scribe, a certain Bi Xianzhi read the text, and its owner was a Buddhist monk, Huiming, from a well-known Waguan monastery located in todays Nanjing. Waguan monastery was first founded in 364 and was still in use during the Liang dynasty. Huiming was likely an influential monk. A few years before the copying of this text, monk Huiming from Waguan monastery collaborated with monk Fayun from Guangzhai monastery and others to preach on the need to stop eating meat during the Tianjian era.
Cave 17 at the Dunhuang cave complex is a cache of religious texts and social records sealed at the beginning of the eleventh century and discovered in 1900. Legal documents from this cave show that ordained Buddhist monks and nuns in Dunhuang actively engaged in legal practices as plaintiffs, defendants, witnesses, mediators, and representatives in the trial of various disputes. Where did these monks and nuns get their legal education? And what roles did they play in circulating legal knowledge in Dunhuang? Addressing these questions is crucial for my larger project on the interaction between religion and law in China as they can help paint a fuller picture of how Buddhism and the state interacted in the legal arena in pre-modern China. Elsewhere, I have discussed how the jurisdictional boundary between the state and the Buddhist monastic community shifted in central China and analyzed cases in which individual monks and nuns litigated in the lay courts in Dunhuang. This article aims to understand the legal landscape on the frontier in Dunhuang, with a particular focus on ordained Buddhists’ access to legal texts and the roles they played in producing and preserving them. My primary sources are colophons on the Chinese manuscripts of Buddhist monastic law texts (Vinaya) discovered in the Dunhuang cave 17. My examination of the said colophons will reveal that ordained Buddhist monks and nuns played various roles in the circulation of legal knowledge in Dunhuang. They were patrons, scribes, and owners of various Buddhist monastic law texts circulated in Dunhuang. In cave 17, we can find Buddhist monastic law texts of all the major surviving traditions. The majority of them are law texts transmitted from the Dharmaguptaka tradition, but those of the Mahāsāṃghika, Sarvāstivāda, Mahīśāsaka, and Mūlasarvāstivāda traditions have also appeared in the same cave. These monastic law texts cover Buddhist liturgical and disciplinary guidelines ranging from the five precepts for the lay Buddhist householder, to the eight temporary fasting precepts, the 10 precepts for ordained novices, the six additional precepts for probationary Buddhist nuns, and the full set of disciplinary rules for fully ordained Buddhist monks and nuns. This examination, however, is not comprehensive. Ideally, we need to combine this examination with a study on the Buddhist monastic law texts written in Tibetan and other languages found in cave 17. Since Shayne Clarke and his team at McMaster University are currently working on a project surveying the Tibetan Vinaya texts from the Dunhuang manuscript collection, I did not discuss the Tibetan materials in the present article. It is also important to note that while my discussion below focuses on Buddhist monastic law texts, those were not the only type of legal texts circulated in Dunhuang. In this cave, we also find fragments of the primary legal texts promulgated by the Tang court. These include code, statutes, regulations, and ordinances. Several judicial judgments from the Tang, one of which I will discuss below, were also found in Dunhuang. Moreover, we have also found damaged copies of the section on The Penal Code from The Book of the Han. Before discussing how ordained Buddhists engaged in the copying, collecting, and preserving of Buddhist monastic law texts in Dunhuang, it is necessary to provide an overview of the workflow of text copying in Dunhuang. Scribes in Dunhuang copied Buddhist legal texts over several centuries. The earliest Buddhist monastic law text found in cave 17 is the Sarvāstivāda Prātimokṣa for Ordained Buddhist Monks . The scribe, a Buddhist monk Deyou, stated in the colophon that he copied this text in the southern part of Dunhuang during the summer rainy season retreat in 406, one year after he received his full ordination in 405. Copying of Buddhist monastic law texts continued in Dunhuang without interruption toward the end of the tenth century. Even after Dunhuang fell under Tibetan control, Buddhist text copying continued from 786 to 848. The Tibetan scholar Zhang Yanqing’s extensive study on the copying of Tibetan Buddhist texts in Dunhuang helps unveil important aspects of the workflow of text copying in this area. At the beginning of the Tibetan occupation of Dunhuang, text-copying focused on Chinese Buddhist texts. With support from the Tibetan king Khri gtsug lde brtsan, massive copying of Tibetan Buddhist texts began in 822 and continued in Shazhou until 838. Teams stationed at a particular scriptorium (gur) with multiple scribes and proofreaders copied large texts. Before starting the copying, they would secure a mother copy (dpe’) and proofread it multiple times. Each scribe would receive a mother copy of their targeted text and paper, and an administrator would record the distribution of paper and mother copies. A loan record on Pelliot chinois 3010v illustrates this practice. This manuscript states that on the seventeenth day in the fifth month in a mouse year , Sanjie monastery, one of the local monasteries in Dunhuang, borrowed six bundles of the Chinese translation of the Dharmaguptaka Vinaya and their brocade bundle covers for copying . When a scribe completed a copy, a proofreader would check the replica against the original. It was normal to proofread the copied Tibetan Buddhist texts at least three or four times, or 10 times in rare cases. In contrast to proofreading the Buddhist text commissioned by the Tibetan king anywhere between three times and 10 times, proofreaders normally proofread the Chinese copies of Buddhist monastic law texts found in Dunhuang once, twice, or none. For instance, Jingtu monastery owned a copy of the Commentary on the Dharmaguptaka Prātimokṣa. Its colophon informs us that an unspecified proofreader proofread it once at the Mogao cave on the first day in the third month in a horse year. Copies of the Buddhist monastic law texts found in cave 17 were produced in various locations. Some of them were locally copied in Dunhuang. On the nineteenth day in the eleventh month in a dog year, an unspecified scribe copied A Short Collection of Buddhist Monastic Law Texts at Bao’en monastery in Dunhuang. On the twenty-eighth day in the third month in a mouse year, another unspecified scribe copied the third scroll of the Commentary on the Dharmaguptaka Prātimokṣa at Yongshou monastery in Shazhou. This copy was proofread and punctuated in red ink. On the twentieth day in the tenth month in the seventh year of a certain era, a Buddhist monk, Riding , copied The Dharmaguptaka Prātimokṣa for Ordained Buddhist Monks at Dayun monastery. In the twelfth month in a mouse year, a monk from Jinguangming monastery, who was also a Vinaya master, copied the Commentary on the Dharmaguptaka Prātimokṣa for circulation. On the twentieth day in the fourth month in the second year during the Qianyuan era (759), monk Jingshen from Longxing monastery copied the Dharmaguptaka Prātimokṣa for Ordained Buddhist Monks. Another Longxing monastery monk, Zhizhao, copied the Dharmaguptaka Prātimokṣa for Ordained Buddhist Nuns. From the eighth day in the fifth month to the third day in the sixth month in a horse year, a Buddhist monk, Liji, from Jinguangming monastery copied one scroll of the Tang Buddhist monk Daoxuans (596–667) Randomly Selected Ecclesiastical Proceedings: Abridgements and Emendations to the Dharmaguptaka Vinaya for his monastery’s elder monk, Jinyao, at the beginning of that year’s rainy season retreat. Another local site where multiple Buddhist legal texts were copied is the Dongshan solitary monastic dwelling, which was most likely one of many such dwellings built near the Dunhuang cave complex. At least three surviving Buddhist law texts were copied at this site in a tiger year. On the sixteenth day in the sixth month in a tiger year, an unspecified individual completed proofreading the Notes on Conduct: Abridgements and Emendations to the Dharmaguptaka Vinaya at the Dongshan solitary monastic dwelling. One month later, on the fifteenth day in the seventh month of a tiger year, an unspecified scribe finished copying a single-scroll text, A Concise Copy of the Vinaya at Dongshan, identifiable with the Dongshan solitary monastic dwelling mentioned above. The owner of this copied text was a Buddhist nun, Shengzang. A few months later, on the twentieth day in the tenth month in a tiger year, an unspecified scribe continued copying and proofreading the fourth scroll of the Commentary on the Dharmaguptaka Prātimokṣa. A bit further away from Dunhuang, the Xiuduo monastery in Ganzhou Prefecture in todays Gansu province was also an important site for the circulation of legal and other Buddhist knowledge in Dunhuang. A Buddhist monk, Huizhen, copied The Sūtra on the Twenty-four Precepts for the Faithful Bodhisattvas Told by the Buddha at this Xiuduo monastery. The date was unspecified. This monastery was also where the Tibetan Buddhist monk, Chos grub (Ch. Facheng), taught and translated texts. Turfan, located about 800 kilometers to the Northwest of Dunhuang, appeared to be another productive center for copying Buddhist law texts. On the fourth day in the seventh month in the second year during the Guangde era, a Buddhist monk, Yilin, copied Daoxuans Notes on Conduct: Abridgements and Emendations to the Dharmaguptaka Vinaya at a place specified as Peijiata in the Nanping city in the Xizhou Prefecture, in today’s Turfan. At the end of the third and last scroll of this text, the colophons mention the scribe’s name, where and when he copied it. At the end of the middle scroll of this text, an unspecified proofreader wrote in red ink that he proofread the copied text in the early half of the sixth month in a snake year at the garden of a central Asian mound, whose exact location remains unknown. Because the nearest snake year after 764 was 765, this text was likely proofread in 765. Some of the Buddhist monastic law texts traveled to Dunhuang from central China. In the Dunhuang cave 17, we found at least two Buddhist monastic law texts previously owned by ordained Buddhists in central China. One is The Manual for Ordained Buddhists to Receive the Bodhisattva Precepts, which was copied under imperial order in the fifth month in the eighteenth year during the Tianjian era (519) during the Liang Dynasty. A layperson, Dai Mengtong, was the scribe, a certain Bi Xianzhi read the text, and its owner was a Buddhist monk, Huiming, from a well-known Waguan monastery located in todays Nanjing. Waguan monastery was first founded in 364 and was still in use during the Liang dynasty. Huiming was likely an influential monk. A few years before the copying of this text, monk Huiming from Waguan monastery collaborated with monk Fayun from Guangzhai monastery and others to preach on the need to stop eating meat during the Tianjian era.
考点1:“Sarvāstivāda Prātimokṣa for Ordained Buddhist Monks”推荐译为“《十诵律比丘戒本》” 考点2:“bundle”必须译为“帙” 考点3:“Dharmaguptaka Vinaya”必须译为“《四分律》” 考点4:“Randomly Selected Ecclesiastical Proceedings: Abridgements and Emendations to the Dharmaguptaka Vinaya”必须译为“《四分律删补随机羯磨》” 考点5:“Notes on Conduct: Abridgements and Emendations to the Dharmaguptaka Vinaya”必须译为“《四分律删繁补阙行事钞》” 考点6:“solitary monastic dwelling”必须译为“兰若” 考点7:“in red ink”必须译为“朱笔”
51f04
学术论文
人文科学
133
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 空间智能作为人工智能领域的前沿分支,其核心在于构建机器对三维物理世界的结构化认知与动态响应能力。这一概念突破了传统计算机视觉的平面局限,通过整合深度感知、几何推理与运动规划,实现从视觉输入到物理操作的完整闭环。在技术层面,空间智能的本质是通过多模态传感器数据解析三维拓扑结构、语义关联与动态变化规律,使智能体具备类人级的环境理解与行为决策能力。相较于二维视觉仅能捕捉色彩与纹理信息,空间智能通过深度估计、立体匹配与运动恢复结构等技术路径,构建包含度量尺度、空间方位与材质属性的三维表征体系,为自动驾驶、机器人操作等复杂任务提供认知基础。 空间智能的发展历程呈现出从确定性几何建模向概率性神经感知的范式转变。早期探索聚焦于基于刚体运动假设的几何约束方法,其中同步定位与地图构建技术通过迭代优化传感器姿态与环境特征的空间关系,实现未知环境中的自主导航。该技术依赖扩展卡尔曼滤波或 Bundle Adjustment 等优化框架,通过最小化重投影误差实现状态估计,但在动态场景与特征缺失环境中易出现漂移现象。视觉里程计作为 SLAM 的前端模块,通过相邻帧特征点匹配计算运动变换矩阵,其精度直接取决于特征提取的鲁棒性 ——SIFT 与 ORB 等局部特征 descriptor 在光照变化与运动模糊场景中的性能差异,成为制约早期系统稳定性的关键因素。 针对三维数据的稀疏性与高维度特性,NVIDIA 提出的 DeepVDB 框架基于体素化稀疏数据结构,通过层级化存储与索引机制,实现大规模三维卷积与注意力运算的高效加速。该框架在城市级数字孪生构建中,将 TB 级点云数据压缩比提升 30 倍,同时保持亚毫米级几何精度,为实时渲染与物理仿真提供支撑。多模态融合技术则通过摄像头、激光雷达与惯性测量单元的时空校准,构建冗余感知体系:视觉传感器提供丰富纹理信息,激光雷达保证恶劣天气下的测距精度,惯性测量单元则通过捷联惯导算法弥补传感器采样间隙,三者的紧耦合融合通过卡尔曼滤波或图优化实现状态互补,显著提升复杂环境下的感知鲁棒性。 自动驾驶作为空间智能最成熟的应用场景,正经历从多传感器融合向端到端学习的范式转变。当最新研究表明,随着算力投入与训练数据规模的指数级增长,自动驾驶系统呈现显著的缩放定律 — 当训练里程超过 1000 万公里时,系统决策错误率随算力扩展呈幂律下降。特斯拉纯视觉方案基于 Transformer 架构的 BEV 空间特征融合,通过时序注意力机制聚合多帧图像信息,在城市道路场景中实现厘米级定位精度;而 Waymo 的激光雷达方案则通过高精地图与实时点云配准,在复杂路口场景中保持 99.9% 的障碍物识别准确率。尽管技术路线存在差异,但均通过强化学习优化决策策略,使系统在无保护左转、紧急避让等复杂场景中的表现逐步接近人类驾驶员水平。 三维生成与数字孪生技术正推动虚拟世界构建从手工建模向自动化生成的转变。神经辐射场通过多层感知器参数化场景的辐射特性与密度分布,从二维图像集合中重建具有连续视图一致性的三维场景,其核心在于体积渲染方程的数值求解与位置编码机制对高频特征的捕捉。Mip-NeRF 通过锥形采样解决传统 NeRF 的模糊问题,将重建精度提升 40%;Instant NeRF 则引入多分辨率哈希编码,将训练时间从数天压缩至分钟级。这些技术突破使数字孪生从概念走向实用:在城市规划领域,通过机载激光雷达与航空影像融合构建的数字孪生模型,可精确模拟建筑群的日照阴影与风场分布;在工业制造中,基于机器视觉与机器人扫描的设备数字孪生,能实现毫米级的装配误差检测。目前该领域已积累超过 1000 万的三维模型数据,但存在表示形式碎片化问题 — 点云、网格、体素与隐式表示各有优劣,缺乏统一的标准格式,导致跨平台模型复用率不足 30%。 具身智能作为空间智能的终极形态,致力于实现机器人与物理世界的深度交互。与传统工业机器人的预编程控制不同,具身智能系统通过视觉 - 触觉 - 力觉多模态感知,在未知环境中自主完成操作任务。最新研究采用深度强化学习训练机器人操作策略:通过触觉传感器获取物体表面纹理与接触力信息,结合视觉图像的目标姿态估计,在虚拟环境中进行千万次试错训练,再通过域适应技术迁移至物理世界。谷歌的 SayCan 系统将自然语言指令与视觉感知映射至机器人动作空间,在厨房场景中实现餐具整理、食材处理等复杂任务的零样本迁移;波士顿动力的 Atlas 机器人则通过全身运动规划与动态平衡控制,在野外地形中完成跳跃、攀爬等动态操作。尽管取得显著进展,具身智能仍面临诸多挑战:物理世界的无限状态空间导致训练数据采集成本高昂,机器人操作的安全性约束限制探索空间,而动态环境中的物体交互动力学建模仍是未解决的难题。当前研究正通过构建家庭与工业环境的数据闭环系统,实现机器人行为的持续评估与强化学习更新,逐步提升系统的环境适应能力。 斯坦福大学李飞飞团队首次系统提出空间智能的理论框架,强调其核心在于构建 "感知 - 推理 - 执行" 的闭环认知体系。该团队开发的 SceneScape 模型通过单目图像的深度估计与语义分割,结合先验知识图谱推理物体空间关系,从单张图片中生成包含物理属性的完整三维场景,其创新点在于引入因果推理机制处理物体遮挡与空间布局歧义问题,使生成场景的物理合理性提升 65%。World Labs 公司基于该理论开发的商用模型,已在建筑设计领域实现从手绘草图到三维建筑模型的自动转换,设计效率提升 3 倍以上。 清华大学提出的 SpatialBot 大模型,通过 Transformer 架构统一处理点云、图像与文本信息,在三维目标检测任务中实现 89.7% 的 mAP 值;北京大学团队开发的多模态空间注意力网络,创新性地将语言指令编码为空间注意力掩码,使机器人抓取成功率在未知物体上提升至 92%。企业层面,华为发布的 AR Engine 4.0 通过 SLAM 与环境光适应技术,实现手机端厘米级的空间定位精度,支撑虚实融合的导航与教育应用;百度 Apollo 的空间感知系统采用激光雷达与摄像头的紧耦合融合,在城市道路测试中实现 99.99% 的定位可用性;腾讯的数字孪生平台则整合游戏引擎与 AI 生成技术,构建的虚拟城市可实时模拟 10 万级人口的动态行为。 扩展现实作为空间智能的原生交互界面,正重构人机交互范式。虚拟现实通过头显设备构建完全沉浸式的虚拟空间,其空间定位依赖 Inside-out 与 Outside-in 混合追踪技术 —— 前者通过头显内置的视觉惯性传感器实现六自由度定位,后者则通过外部基站的激光扫描提供亚毫米级精度。增强现实则通过光学透视眼镜将虚拟信息叠加于物理世界,微软 HoloLens 2 采用多层衍射波导与时间复用技术,实现明亮环境下的高对比度显示,其空间映射功能可实时构建房间尺度的三维网格。这些技术突破使 XR 成为空间智能的天然载体:在远程协作场景中,借助 XR 设备的空间感知能力,异地用户可共享虚拟工作台并进行三维物体的协同操作;在技能培训领域,基于空间智能的 XR 模拟器能精确还原手术器械或工业设备的操作手感与物理反馈。 未来空间智能将通过与多学科技术的深度融合实现认知跃迁。与大语言模型的结合将赋予系统理解空间描述语言的能力,使 "将沙发移至窗户左侧" 等自然指令转化为精确的空间坐标与运动轨迹;机器人控制领域的阻抗控制与力位混合控制技术,将提升空间智能的物理执行精度,实现精密装配等微米级操作;脑机接口则可能成为终极交互方式,通过解析大脑运动皮层的空间意图信号,直接控制外部设备的三维运动。这些技术融合将最终实现从 "视觉感知" 到 "物理行动" 的完整闭环,使人工智能系统具备与人类相当的空间认知与操作能力,彻底重塑制造业、医疗健康、智能家居等领域的技术形态。
空间智能作为人工智能领域的前沿分支,其核心在于构建机器对三维物理世界的结构化认知与动态响应能力。这一概念突破了传统计算机视觉的平面局限,通过整合深度感知、几何推理与运动规划,实现从视觉输入到物理操作的完整闭环。在技术层面,空间智能的本质是通过多模态传感器数据解析三维拓扑结构、语义关联与动态变化规律,使智能体具备类人级的环境理解与行为决策能力。相较于二维视觉仅能捕捉色彩与纹理信息,空间智能通过深度估计、立体匹配与运动恢复结构等技术路径,构建包含度量尺度、空间方位与材质属性的三维表征体系,为自动驾驶、机器人操作等复杂任务提供认知基础。 空间智能的发展历程呈现出从确定性几何建模向概率性神经感知的范式转变。早期探索聚焦于基于刚体运动假设的几何约束方法,其中同步定位与地图构建技术通过迭代优化传感器姿态与环境特征的空间关系,实现未知环境中的自主导航。该技术依赖扩展卡尔曼滤波或 Bundle Adjustment 等优化框架,通过最小化重投影误差实现状态估计,但在动态场景与特征缺失环境中易出现漂移现象。视觉里程计作为 SLAM 的前端模块,通过相邻帧特征点匹配计算运动变换矩阵,其精度直接取决于特征提取的鲁棒性 ——SIFT 与 ORB 等局部特征 descriptor 在光照变化与运动模糊场景中的性能差异,成为制约早期系统稳定性的关键因素。 针对三维数据的稀疏性与高维度特性,NVIDIA 提出的 DeepVDB 框架基于体素化稀疏数据结构,通过层级化存储与索引机制,实现大规模三维卷积与注意力运算的高效加速。该框架在城市级数字孪生构建中,将 TB 级点云数据压缩比提升 30 倍,同时保持亚毫米级几何精度,为实时渲染与物理仿真提供支撑。多模态融合技术则通过摄像头、激光雷达与惯性测量单元的时空校准,构建冗余感知体系:视觉传感器提供丰富纹理信息,激光雷达保证恶劣天气下的测距精度,惯性测量单元则通过捷联惯导算法弥补传感器采样间隙,三者的紧耦合融合通过卡尔曼滤波或图优化实现状态互补,显著提升复杂环境下的感知鲁棒性。 自动驾驶作为空间智能最成熟的应用场景,正经历从多传感器融合向端到端学习的范式转变。当最新研究表明,随着算力投入与训练数据规模的指数级增长,自动驾驶系统呈现显著的缩放定律 — 当训练里程超过 1000 万公里时,系统决策错误率随算力扩展呈幂律下降。特斯拉纯视觉方案基于 Transformer 架构的 BEV 空间特征融合,通过时序注意力机制聚合多帧图像信息,在城市道路场景中实现厘米级定位精度;而 Waymo 的激光雷达方案则通过高精地图与实时点云配准,在复杂路口场景中保持 99.9% 的障碍物识别准确率。尽管技术路线存在差异,但均通过强化学习优化决策策略,使系统在无保护左转、紧急避让等复杂场景中的表现逐步接近人类驾驶员水平。 三维生成与数字孪生技术正推动虚拟世界构建从手工建模向自动化生成的转变。神经辐射场通过多层感知器参数化场景的辐射特性与密度分布,从二维图像集合中重建具有连续视图一致性的三维场景,其核心在于体积渲染方程的数值求解与位置编码机制对高频特征的捕捉。Mip-NeRF 通过锥形采样解决传统 NeRF 的模糊问题,将重建精度提升 40%;Instant NeRF 则引入多分辨率哈希编码,将训练时间从数天压缩至分钟级。这些技术突破使数字孪生从概念走向实用:在城市规划领域,通过机载激光雷达与航空影像融合构建的数字孪生模型,可精确模拟建筑群的日照阴影与风场分布;在工业制造中,基于机器视觉与机器人扫描的设备数字孪生,能实现毫米级的装配误差检测。目前该领域已积累超过 1000 万的三维模型数据,但存在表示形式碎片化问题 — 点云、网格、体素与隐式表示各有优劣,缺乏统一的标准格式,导致跨平台模型复用率不足 30%。 具身智能作为空间智能的终极形态,致力于实现机器人与物理世界的深度交互。与传统工业机器人的预编程控制不同,具身智能系统通过视觉 - 触觉 - 力觉多模态感知,在未知环境中自主完成操作任务。最新研究采用深度强化学习训练机器人操作策略:通过触觉传感器获取物体表面纹理与接触力信息,结合视觉图像的目标姿态估计,在虚拟环境中进行千万次试错训练,再通过域适应技术迁移至物理世界。谷歌的 SayCan 系统将自然语言指令与视觉感知映射至机器人动作空间,在厨房场景中实现餐具整理、食材处理等复杂任务的零样本迁移;波士顿动力的 Atlas 机器人则通过全身运动规划与动态平衡控制,在野外地形中完成跳跃、攀爬等动态操作。尽管取得显著进展,具身智能仍面临诸多挑战:物理世界的无限状态空间导致训练数据采集成本高昂,机器人操作的安全性约束限制探索空间,而动态环境中的物体交互动力学建模仍是未解决的难题。当前研究正通过构建家庭与工业环境的数据闭环系统,实现机器人行为的持续评估与强化学习更新,逐步提升系统的环境适应能力。 斯坦福大学李飞飞团队首次系统提出空间智能的理论框架,强调其核心在于构建 "感知 - 推理 - 执行" 的闭环认知体系。该团队开发的 SceneScape 模型通过单目图像的深度估计与语义分割,结合先验知识图谱推理物体空间关系,从单张图片中生成包含物理属性的完整三维场景,其创新点在于引入因果推理机制处理物体遮挡与空间布局歧义问题,使生成场景的物理合理性提升 65%。World Labs 公司基于该理论开发的商用模型,已在建筑设计领域实现从手绘草图到三维建筑模型的自动转换,设计效率提升 3 倍以上。 清华大学提出的 SpatialBot 大模型,通过 Transformer 架构统一处理点云、图像与文本信息,在三维目标检测任务中实现 89.7% 的 mAP 值;北京大学团队开发的多模态空间注意力网络,创新性地将语言指令编码为空间注意力掩码,使机器人抓取成功率在未知物体上提升至 92%。企业层面,华为发布的 AR Engine 4.0 通过 SLAM 与环境光适应技术,实现手机端厘米级的空间定位精度,支撑虚实融合的导航与教育应用;百度 Apollo 的空间感知系统采用激光雷达与摄像头的紧耦合融合,在城市道路测试中实现 99.99% 的定位可用性;腾讯的数字孪生平台则整合游戏引擎与 AI 生成技术,构建的虚拟城市可实时模拟 10 万级人口的动态行为。 扩展现实作为空间智能的原生交互界面,正重构人机交互范式。虚拟现实通过头显设备构建完全沉浸式的虚拟空间,其空间定位依赖 Inside-out 与 Outside-in 混合追踪技术 —— 前者通过头显内置的视觉惯性传感器实现六自由度定位,后者则通过外部基站的激光扫描提供亚毫米级精度。增强现实则通过光学透视眼镜将虚拟信息叠加于物理世界,微软 HoloLens 2 采用多层衍射波导与时间复用技术,实现明亮环境下的高对比度显示,其空间映射功能可实时构建房间尺度的三维网格。这些技术突破使 XR 成为空间智能的天然载体:在远程协作场景中,借助 XR 设备的空间感知能力,异地用户可共享虚拟工作台并进行三维物体的协同操作;在技能培训领域,基于空间智能的 XR 模拟器能精确还原手术器械或工业设备的操作手感与物理反馈。 未来空间智能将通过与多学科技术的深度融合实现认知跃迁。与大语言模型的结合将赋予系统理解空间描述语言的能力,使 "将沙发移至窗户左侧" 等自然指令转化为精确的空间坐标与运动轨迹;机器人控制领域的阻抗控制与力位混合控制技术,将提升空间智能的物理执行精度,实现精密装配等微米级操作;脑机接口则可能成为终极交互方式,通过解析大脑运动皮层的空间意图信号,直接控制外部设备的三维运动。这些技术融合将最终实现从 "视觉感知" 到 "物理行动" 的完整闭环,使人工智能系统具备与人类相当的空间认知与操作能力,彻底重塑制造业、医疗健康、智能家居等领域的技术形态。
考点1:“BEV空间特征融合” 推荐译为 BEV spatial feature fusion 考点2:“幂律缩放定律” 推荐译为scaling laws 考点3:“时序注意力融合” 推荐译为 temporal attention fusion 考点4:“视觉 - 触觉 - 力觉多模态感知” 推荐译为 vision‑tactile‑force multimodal perception
51fa8
学术论文
应用学科
176
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 神圣空间受到了来自不同学科的学者的关注,这些不同的研究扩充了神圣空间的概念。这些定义可以大致可以分为两类:实质性的和情境性的。前者从内在的视角出发,说明神圣空间的本质特征为令人敬畏的、不可思议的和令人不安的。后者从外在的视角出发,将神圣空间定位在人类实践与社会活动的交叉点,并将其理论化为“空的能指” 。基于这些定义,神圣空间的研究对象从传统的宗教空间和祭祀空间扩展开来,学者开始关注盒子里的教堂、朝圣城市、人口稀少的农村地区的临时神圣空间和纪念碑等非传统神圣空间。“神圣”被理解为是人类行动的产物,因此任何空间都有可能成为神圣空间。 孔庙是中国一种历史悠久的坛庙建筑。孔子(前551-前479),山东曲阜人,是中国古代伟大的思想家、政治家、教育家,儒家学派创始人。自汉武帝罢黜百家独尊儒术后,历代帝王多以儒学作为指导思想,孔子地位日崇,孔庙的规模也日益宏大壮丽。汉高祖十二年(前195)以太牢祀孔子庙堂,开创了帝王祭祀孔子的先例。唐贞观四年(630年),唐太宗下诏在京城和全国各地设立孔庙。唐开元二十六年(739年),《唐六典》成书,它是中国目前保存的最早的、最完整的、古代国家行政法典《唐六典》成书,该书将祭祀孔子列为于国家祭祀典礼之中。此书颁布是,宣告了孔庙祭祀制度的正式定型的里程碑 。此后孔庙的建设和祭孔礼仪规格不断扩大,各地文庙建设活动均在明清时达到顶峰。孔庙除了遍布中国外,还扩散到日本、韩国和东南亚许多国家。 中国古代坛庙主要有三类:祭祀自然神(如天坛、地坛、先农坛等)、祭祀祖先(如太庙、家庙等)以及祭祀先贤(如武侯祠、关帝庙等)。孔庙兼有第二和第三类,以第三类数量最多。孔氏家庙是孔氏族人祭祀先祖孔子的重要场所,具有特殊地位和特殊内涵。《阙里文献考》记载:先圣殁世,弟子将其葬于鲁城北泗上。殁后二年(前478年),鲁哀公尊夫子为“尼父”,命将其居所改为庙堂,岁时祭奉,此为孔庙之始。中国,目前仅存三座孔氏家庙:山东曲阜孔庙,浙江衢州南宗孔氏家庙、浙江婺州南宗榉溪孔氏家庙。北魏时期,孝文帝于太和十三年(489年)在都城平城(今山西大同)修建孔庙,这是历史上第一座在曲阜以外的孔庙,为之后各地孔庙的建设开创了先例 。由官府修建的孔庙分为三级:国庙、州府庙、县庙。在一些县下面的乡镇也设有孔庙,如云南楚雄大石羊古镇姚县石羊古镇的文庙,这些都是民间修建的。 孔庙与学宫两组建筑毗邻。魏黄初二年(221年),魏文帝于阙里孔庙之外修建了学宫,据说这是可查到的在孔庙旁边修建学宫的最早记录 。唐贞观四年(630年),太宗下命令在全国各州县的学宫里设立孔庙。学宫是各级政府在其所在城市里设立的最高等级的学府,相当于在首都的最高学府。 唐代后期,孔庙从学宫中分离开,成为独立的建筑群。至宋代,孔庙和学宫基本上成为毗邻的两组独立建筑群。如今在中国许多古城中,人们都可以看到孔庙和学宫两个相邻的建筑群。通常是站在学宫里面对大门,其左边的建筑群为孔庙,这是因为在中国传统中,左边的等级比右边高 。如图1 清代北京(当时国家的首都)下辖大兴县的学宫和孔庙建筑群平面图。也有例外,如顺天府的孔庙图所示(图2)。少数孔庙建筑群是在学宫的前面。 孔庙经过2000多年的发展演变,到明代三级官府修建的孔庙基本形成了稳定的建筑群构成。从图1可以看到,建筑的中轴线上,从南到北依次是照壁;庙门;泮池,为一个半月形的水池,上面有一座桥连接中轴线上的主路;大成门,是进入主院的入口;大成殿,是供奉孔子牌位,或孔子塑像的主殿,有的还在孔子塑像两侧,供奉孔子的两位优秀的学生塑像 。“大成”的中文含义是“所有智慧的集合”,这个词来自孔子思想的追随者孟子。孟子将孔子称为圣人。因为孔子将所有前人智慧汇集在一起,所以供奉他的殿宇就称为“大成殿”。规模比较大的孔庙还要包括崇圣祠,是供奉孔子父母牌位的殿宇,有的也称启圣祠(见图3右侧最北面的建筑上的汉字)。在主要院落的东西两侧为附属建筑(另加图3右侧的孔庙图)。
神圣空间受到了来自不同学科的学者的关注,这些不同的研究扩充了神圣空间的概念。这些定义可以大致可以分为两类:实质性的和情境性的。前者从内在的视角出发,说明神圣空间的本质特征为令人敬畏的、不可思议的和令人不安的。后者从外在的视角出发,将神圣空间定位在人类实践与社会活动的交叉点,并将其理论化为“空的能指” 。基于这些定义,神圣空间的研究对象从传统的宗教空间和祭祀空间扩展开来,学者开始关注盒子里的教堂、朝圣城市、人口稀少的农村地区的临时神圣空间和纪念碑等非传统神圣空间。“神圣”被理解为是人类行动的产物,因此任何空间都有可能成为神圣空间。 孔庙是中国一种历史悠久的坛庙建筑。孔子(前551-前479),山东曲阜人,是中国古代伟大的思想家、政治家、教育家,儒家学派创始人。自汉武帝罢黜百家独尊儒术后,历代帝王多以儒学作为指导思想,孔子地位日崇,孔庙的规模也日益宏大壮丽。汉高祖十二年(前195)以太牢祀孔子庙堂,开创了帝王祭祀孔子的先例。唐贞观四年(630年),唐太宗下诏在京城和全国各地设立孔庙。唐开元二十六年(739年),《唐六典》成书,它是中国目前保存的最早的、最完整的、古代国家行政法典《唐六典》成书,该书将祭祀孔子列为于国家祭祀典礼之中。此书颁布是,宣告了孔庙祭祀制度的正式定型的里程碑 。此后孔庙的建设和祭孔礼仪规格不断扩大,各地文庙建设活动均在明清时达到顶峰。孔庙除了遍布中国外,还扩散到日本、韩国和东南亚许多国家。 中国古代坛庙主要有三类:祭祀自然神(如天坛、地坛、先农坛等)、祭祀祖先(如太庙、家庙等)以及祭祀先贤(如武侯祠、关帝庙等)。孔庙兼有第二和第三类,以第三类数量最多。孔氏家庙是孔氏族人祭祀先祖孔子的重要场所,具有特殊地位和特殊内涵。《阙里文献考》记载:先圣殁世,弟子将其葬于鲁城北泗上。殁后二年(前478年),鲁哀公尊夫子为“尼父”,命将其居所改为庙堂,岁时祭奉,此为孔庙之始。中国,目前仅存三座孔氏家庙:山东曲阜孔庙,浙江衢州南宗孔氏家庙、浙江婺州南宗榉溪孔氏家庙。北魏时期,孝文帝于太和十三年(489年)在都城平城(今山西大同)修建孔庙,这是历史上第一座在曲阜以外的孔庙,为之后各地孔庙的建设开创了先例 。由官府修建的孔庙分为三级:国庙、州府庙、县庙。在一些县下面的乡镇也设有孔庙,如云南楚雄大石羊古镇姚县石羊古镇的文庙,这些都是民间修建的。 孔庙与学宫两组建筑毗邻。魏黄初二年(221年),魏文帝于阙里孔庙之外修建了学宫,据说这是可查到的在孔庙旁边修建学宫的最早记录 。唐贞观四年(630年),太宗下命令在全国各州县的学宫里设立孔庙。学宫是各级政府在其所在城市里设立的最高等级的学府,相当于在首都的最高学府。 唐代后期,孔庙从学宫中分离开,成为独立的建筑群。至宋代,孔庙和学宫基本上成为毗邻的两组独立建筑群。如今在中国许多古城中,人们都可以看到孔庙和学宫两个相邻的建筑群。通常是站在学宫里面对大门,其左边的建筑群为孔庙,这是因为在中国传统中,左边的等级比右边高 。如图1 清代北京(当时国家的首都)下辖大兴县的学宫和孔庙建筑群平面图。也有例外,如顺天府的孔庙图所示(图2)。少数孔庙建筑群是在学宫的前面。 孔庙经过2000多年的发展演变,到明代三级官府修建的孔庙基本形成了稳定的建筑群构成。从图1可以看到,建筑的中轴线上,从南到北依次是照壁;庙门;泮池,为一个半月形的水池,上面有一座桥连接中轴线上的主路;大成门,是进入主院的入口;大成殿,是供奉孔子牌位,或孔子塑像的主殿,有的还在孔子塑像两侧,供奉孔子的两位优秀的学生塑像 。“大成”的中文含义是“所有智慧的集合”,这个词来自孔子思想的追随者孟子。孟子将孔子称为圣人。因为孔子将所有前人智慧汇集在一起,所以供奉他的殿宇就称为“大成殿”。规模比较大的孔庙还要包括崇圣祠,是供奉孔子父母牌位的殿宇,有的也称启圣祠(见图3右侧最北面的建筑上的汉字)。在主要院落的东西两侧为附属建筑(另加图3右侧的孔庙图)。
考点1:“太牢”必须翻译为“sacrificial ox”,因为清朝以后太牢祭祀才祭祀羊和猪,在此之前只祭祀牛,原文说的是清朝前而非清朝后。 考点2:“唐六典”必须翻译为“Tang Liudian”,因为这是出版物译法 考点3:“天坛”必须翻译为“the Temple of Heaven”,景点固定译法 考点4:“地坛”必须翻译为“Temple of Earth”,景点固定译法 考点5:“先农坛”必须翻译为“Xiannong Altar”,北京旅游网站固定译法 考点6:“太庙”必须翻译为“the Imperial Ancestral Temple”,景点固定译法 考点7:“家庙”推荐翻译为“clan temple” 考点8:“武侯祠”必须译为“Wuhou Shrine”,景点固定译法 考点9:“阙里文献考”必须译为“Study of Queli documents”,出版物译法 考点10:“泮池”推荐翻译为“the Pan Chi” 考点11:“大成门”必须译为“the Dacheng Gate”,景点固定译法 考点12:“崇圣祠”必须译为“Chongsheng Shrine”,景点固定译法
583a5
学术论文
人文科学
161
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 社会主义市场经济制度优势可以实现“供给约束”突破通缩困局,调控行业供需平衡来规避经济下行风险。社会主义市场经济以国有经济为主体,提供强大的宏观调控能力,有能力实现避免资本主义市场经济周期波动。面对通货紧缩,传统凯恩斯主义的需求刺激政策因悲观预期固化而失效,如日本和欧洲案例所示,通缩环境难以通过需求刺激改善经济状况。而中国则可以发挥制度优势,通过“供给约束”策略,主动削减过剩产能或者约束产量,保护优质企业免于恶性竞争,这种制度驱动的“理性约束”不是历史的机械复归,并非回归“计划经济”。是对经济重建供需平衡,等待复苏春风的到来。 供给侧改革已经得到验证,我国众多行业已经实现了产业升级,当前需要“供给约束”、“反内卷”来保护经济发展成果。2016年开启的供给侧结构性改革正是“供给约束”的一次成功实战验证,PPI止跌回升,企业扭亏为盈。当前在全球需求萎缩背景下,供给约束通过政策引导和法律工具,统筹各行业的产能产量,避免资源浪费和内卷,保存产业升级成果。历史证明,这不仅能缓冲经济下行,还为发展中国家提供新治理范式。 钢铁行业产能严重过剩,央国企集中度较高利于行政手段实施,“反内卷”政策有望推动钢价修复,盈利抬升。1)钢铁行业供过于求,螺纹钢、线材产能严重过剩:中国钢铁行业2007-2024年国内持续处于供过于求状态,净出口量2024年位于近15年最高水平。2015-2021年中国钢铁企业五大钢材产能利用率在60%-80%之间波动,2021-2025年分品种来看,螺纹钢及线材产能利用率从70%左右下滑至50%左右,严重过剩。2)钢铁行业央国企产量占比显著高于民企,头部集中度较高,预期此次反内卷政策后续或约束5-10%产量:2024年钢铁重点企业中央国企产量超5亿吨,占总产量约63%,民企产量为3亿吨,占总产量约为37%。2024年钢铁重点企业产量前10家企业央国企占比7家,CR10为51%,集中度较高。目前我们预计供给侧改革2.0或约束粗钢产量5%-10%左右,下放指标主要考虑环保及碳排达标程度,我们预计政策落地后螺纹价格有望修复至3500-4000元/吨左右,盈利或增加50-100元/吨左右。3)建议关注3类标的:利润反转、利润稳定估值修复、利润稳定高分红:①利润反转:建议关注柳钢股份、太钢不锈、山东钢铁、三钢闽光、安阳钢铁等,假设在反内卷情况下,吨钢净利100元/吨,测算得出政策落地后年化PE(截至2025年7月15日)分别为5.9/6.4/8.8/10.7/13.3。②利润稳定、估值修复标的:建议关注PE<15倍或PB<1的钢铁企业:新钢股份、首钢股份、宝钢股份、华菱钢铁等。③利润稳定、高股息标的:建议关注鄂尔多斯、友发集团、久立特材、南钢股份、中信特钢等。 工业金属“内卷”的本质是资源端的匮乏与冶炼端的过剩背景下的结构性失衡。1)铜冶炼:行业产能过剩,2018-2024年我国铜冶炼行业产能利用率在70-85%之间浮动,预计2025年铜冶炼上市公司增加产量85万吨,yoy+8%,行业内卷或进一步加剧,截至2025年7月TC/RC分别为-43美元/干吨、-4.3美分/磅,行业持续亏损,我们预计后续“反内卷”政策下加工费有望先恢复至近10年来最低水平——2024年均值水平,TC为11.3美元/干吨,RC为1.1美分/磅。建议关注2类标的:①规模优势企业估值修复:江西铜业、铜陵有色、紫金矿业、中金岭南等;②利润稳定高分红:西部矿业、江西铜业等。2)铅锌冶炼:我们认为铅锌行业的供给侧改革已在路上,后续具备进一步细则落实的政策空间,我们预计2025年下半年锌精矿TC预计将进一步有序回升至4000—4500元/金属吨;进口铅精矿TC预计将回升至-20美元/干吨;建议关注成本优势豫光金铅、自给率较高,具备完整产业链驰宏锌锗。3)氧化铝:氧化铝冶炼具备结构性产能优化空间,老旧产能的清退对行业盈利端的健康发展起到至关重要的作用,我们预计2025年下半年氧化铝价格将处于2800-3400元/吨区间运行,现阶段氧化铝价格的下行已充分计价,企业盈利已趋稳;建议关注有铝土矿权的中国宏桥、中国铝业、天山铝业、印尼建厂南山铝业。4)电解铝:电解铝供给端已于2017至2018年完成供给侧改革,并于今年迅速完成了其需求结构的优化,“反内卷”对其下游需求的影响较小,我们预计2025年下半年电解铝价格将处于20000—21000元/吨区间运行,企业盈利能力有望进一步增强。建议关注具备电解铝产能指标较高的中国宏桥、中国铝业;推荐绿电铝低耗能的云铝股份。
社会主义市场经济制度优势可以实现“供给约束”突破通缩困局,调控行业供需平衡来规避经济下行风险。社会主义市场经济以国有经济为主体,提供强大的宏观调控能力,有能力实现避免资本主义市场经济周期波动。面对通货紧缩,传统凯恩斯主义的需求刺激政策因悲观预期固化而失效,如日本和欧洲案例所示,通缩环境难以通过需求刺激改善经济状况。而中国则可以发挥制度优势,通过“供给约束”策略,主动削减过剩产能或者约束产量,保护优质企业免于恶性竞争,这种制度驱动的“理性约束”不是历史的机械复归,并非回归“计划经济”。是对经济重建供需平衡,等待复苏春风的到来。 供给侧改革已经得到验证,我国众多行业已经实现了产业升级,当前需要“供给约束”、“反内卷”来保护经济发展成果。2016年开启的供给侧结构性改革正是“供给约束”的一次成功实战验证,PPI止跌回升,企业扭亏为盈。当前在全球需求萎缩背景下,供给约束通过政策引导和法律工具,统筹各行业的产能产量,避免资源浪费和内卷,保存产业升级成果。历史证明,这不仅能缓冲经济下行,还为发展中国家提供新治理范式。 钢铁行业产能严重过剩,央国企集中度较高利于行政手段实施,“反内卷”政策有望推动钢价修复,盈利抬升。1)钢铁行业供过于求,螺纹钢、线材产能严重过剩:中国钢铁行业2007-2024年国内持续处于供过于求状态,净出口量2024年位于近15年最高水平。2015-2021年中国钢铁企业五大钢材产能利用率在60%-80%之间波动,2021-2025年分品种来看,螺纹钢及线材产能利用率从70%左右下滑至50%左右,严重过剩。2)钢铁行业央国企产量占比显著高于民企,头部集中度较高,预期此次反内卷政策后续或约束5-10%产量:2024年钢铁重点企业中央国企产量超5亿吨,占总产量约63%,民企产量为3亿吨,占总产量约为37%。2024年钢铁重点企业产量前10家企业央国企占比7家,CR10为51%,集中度较高。目前我们预计供给侧改革2.0或约束粗钢产量5%-10%左右,下放指标主要考虑环保及碳排达标程度,我们预计政策落地后螺纹价格有望修复至3500-4000元/吨左右,盈利或增加50-100元/吨左右。3)建议关注3类标的:利润反转、利润稳定估值修复、利润稳定高分红:①利润反转:建议关注柳钢股份、太钢不锈、山东钢铁、三钢闽光、安阳钢铁等,假设在反内卷情况下,吨钢净利100元/吨,测算得出政策落地后年化PE(截至2025年7月15日)分别为5.9/6.4/8.8/10.7/13.3。②利润稳定、估值修复标的:建议关注PE<15倍或PB<1的钢铁企业:新钢股份、首钢股份、宝钢股份、华菱钢铁等。③利润稳定、高股息标的:建议关注鄂尔多斯、友发集团、久立特材、南钢股份、中信特钢等。 工业金属“内卷”的本质是资源端的匮乏与冶炼端的过剩背景下的结构性失衡。1)铜冶炼:行业产能过剩,2018-2024年我国铜冶炼行业产能利用率在70-85%之间浮动,预计2025年铜冶炼上市公司增加产量85万吨,yoy+8%,行业内卷或进一步加剧,截至2025年7月TC/RC分别为-43美元/干吨、-4.3美分/磅,行业持续亏损,我们预计后续“反内卷”政策下加工费有望先恢复至近10年来最低水平——2024年均值水平,TC为11.3美元/干吨,RC为1.1美分/磅。建议关注2类标的:①规模优势企业估值修复:江西铜业、铜陵有色、紫金矿业、中金岭南等;②利润稳定高分红:西部矿业、江西铜业等。2)铅锌冶炼:我们认为铅锌行业的供给侧改革已在路上,后续具备进一步细则落实的政策空间,我们预计2025年下半年锌精矿TC预计将进一步有序回升至4000—4500元/金属吨;进口铅精矿TC预计将回升至-20美元/干吨;建议关注成本优势豫光金铅、自给率较高,具备完整产业链驰宏锌锗。3)氧化铝:氧化铝冶炼具备结构性产能优化空间,老旧产能的清退对行业盈利端的健康发展起到至关重要的作用,我们预计2025年下半年氧化铝价格将处于2800-3400元/吨区间运行,现阶段氧化铝价格的下行已充分计价,企业盈利已趋稳;建议关注有铝土矿权的中国宏桥、中国铝业、天山铝业、印尼建厂南山铝业。4)电解铝:电解铝供给端已于2017至2018年完成供给侧改革,并于今年迅速完成了其需求结构的优化,“反内卷”对其下游需求的影响较小,我们预计2025年下半年电解铝价格将处于20000—21000元/吨区间运行,企业盈利能力有望进一步增强。建议关注具备电解铝产能指标较高的中国宏桥、中国铝业;推荐绿电铝低耗能的云铝股份。
考点1:复苏春风 推荐译为 economic recovery,原文表达的是经济复苏的意思,“春风”作为无实际意义的修辞可以选择不做直译 考点2:反内卷 推荐译为 anti-rat-race 考点3:中信特钢 必须译为 Citic Pacific Special Steel Group Co.,Ltd.,不可译为“CITIC Metal" 考点4:中金岭南 必须译为 Shenzhen Zhongjin Lingnan Nonfemet Co.Ltd.,不可译为“China Minmetals Corporation”
5b6eb
新闻资讯
观点评论
181
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Synthesis and characterization of BTO photocatalysts As a layered structural two-dimensional material, BTO is characterized by regular stacking of bismuth oxygen [Bi2O2]2+ slabs and perovskite-like [Bi2Ti3O10]2– blocks, in which Bi–O bond can be etched via the interaction of proton and halide ion. Briefly, pristine single-crystal BTO was firstly prepared by a salt-assisted solid-state reaction. Due to the different structural stability of exposed crystal facets, the [Bi2O2]2+ structure of {001} crystal facet and Bi–O bond of {010} crystal facet of BTO can be selectively preferentially etched (Supplementary Fig. 1 and Supplementary Note 1). Also, the Ti–O bond of BTO {001} will prevent further etching of {001} crystal facet after etching [Bi2O2]2+, ultimately leading to the formation of a hollow structure along the edge region. Scanning electron microscopy (SEM) images show the pristine BTO with an evident nanosheet morphology. When HCl etching was conducted, the etched edge gradually became highly visible with the extension of the etching time to 2 h (Fig. 1a and Supplementary Fig. 2). However, the particle size or crystallite size of samples had no significant change, but the obvious difference of specific surface area occurs before and after acid etching (Supplementary Fig. 3 and Supplementary Table 1). When the etching time extended to 4 h, the BTO-4 shows that the nanosheets are broken with impurities due to nucleation of TiO2 (Supplementary Fig. 4). X-ray diffraction (XRD) pattern indicates the same crystal structure between the pristine and the etched BTO within 2 h (Fig. 1b). However, the crystallinity of the etched BTO is gradually weakened with further extension of etching time (Fig. 1c). Unsurprisingly, the over-etched BTO-4 shows a mixed structure of BTO and TiO2 (Supplementary Figs. 5 to 6 and Supplementary Note 2). Similarly, under the same concentration as HCl, the HBr-treated BTO sample shows similar crystal structure and morphology (Supplementary Figs. 7 and 8). The control experiments were performed with HNO3 and H2SO4 treated pristine BTO under the same concentration as HCl. The XRD results indicate the same crystal structure as BTO (Supplementary Fig. 9). And the edge region of BTO has not been etched (Supplementary Fig. 10). Further, an atomic force microscope (AFM) was used to investigate the thickness change before and after etching for different periods. No significant change appears in the overall thickness before and after etching and between the center and edge of the etched BTO (Supplementary Fig. 11). Furthermore, a highly hollow edge structure was observed by the SEM image, and the corresponding elemental mapping reveals the uniform distributions of Bi and Ti elements (Supplementary Fig. 12). Other SEM images, the corresponding elemental mapping, and linear scan elemental analysis are provided in Supplementary Figs. 13–15, and the results proved the uniform distributions of Bi, Ti, and O elements in both the center and the edge region of BTO with and without HCl treatment. Further characterization of the local structure of the etched BTO was carried out under high-resolution transmission electron microscopy (HRTEM). An evident thickness variation contrast occurred between the central and the edge regions of transmission electron microscopy (TEM) image because of the etching effect (Fig. 2a). The HRTEM image exhibits the same lattice orientation and lattice fringes between the edge and the central regions of the etched BTO (Figs. 2b,c), indicating the single crystalline nature of the structure. Furthermore, the same diffraction patterns further confirm the single-crystal structure by selected area electron diffraction (SAED) at different locations (Figs. 2d and 2e). Accordingly, the chemical valence states of elements were also characterized by X-ray photoelectron spectroscopy (XPS). Due to the gradually etched depth of the structure, the chemical states of the surface elements change in a dynamically evolving manner, especially for Ti and Bi. Evidently, as shown in Fig. 2f, the binding energy of Ti 2p gradually shifts towards a higher value, which comes from the increasing proportion of the Ti–O structures exposed on the surface because of the Ti element binding more high-energy O (hydroxyls and water molecules). Due to the change of lattice oxygen (Bi–O and Ti–O) structure on the exposed surface, the binding energy of O 1 s shifts to higher value at the same time (Supplementary Fig. 16a). In contrast, the corresponding Bi 4 f peak slightly shifts downward in binding energy with increasing extent of the acid etching, because the outside high binding energy Bi of [Bi2O2]2+ layer and [Bi2Ti3O10]2– was etched away. Although the binding energies of Bi and Ti evolve, the main structure of samples keeps unchanged during the etching process (Fig. 2g). No Cl element was observed during the etching process, indicating that the nucleation of bismuth oxyhalide was inhibited (Supplementary Fig. 16b). In addition, the structural change should be accompanied by the evolution of the band structure. The band gap slightly increases with the extension of the etching, which is due to the change of the band structure by the spatial size effect, which is mainly from the ultrathin hollow edge (Supplementary Figs. 17–19). With the increase of edge etching depth, the trend of band structure change is more obvious. Meanwhile, the difference in the band structures between the unetched part and adjacent etched BTO will form “homojunction”, which are beneficial to the transfer of electrons to the surface of the photocatalyst. Also, due to TiO2 impurities resulting from structural destruction caused by the excessive etching, the BTO-4 further shows the bigger band gap and evident differences from BTO in light absorption behavior. However, such a small change of bandgaps is not expected to influence the photocatalytic performance. Further understanding the charge distribution on the crystal surface is helpful to explore the process of photocatalytic water splitting. SPVM is an effective tool for spatial correlation between photogenerated charge distribution and structural morphology, as well as the dependence between light absorption and carrier migration and transfer46,47. As shown in Figs. 3a–f, photogenerated electrons were imaged in the surface due to negative SPV. With the increasing wavelength of illumination light, the intensity of the SPV decreases, which is consistent with the behavior of the light absorption spectrum, which proves that the charge generation is driven by light (Supplementary Fig. 17 and Fig. 3g). The difference is that when the light irradiated the surface of photocatalyst, the charge distribution intensity varies between the edge and the middle of the sample. This phenomenon is very interesting and differs from most of the results reported in the literatures. We speculate the carrier distribution is dependent on the space geometry. Differential SPV analysis was performed at different positions of a given facet and the differences were compared at the three positions. The corresponding results are shown in Figs. 3a–f. The SPV is significantly different with the change of light wavelength at the three locations of α1, α2 and α3. Therefore, the intensity of SPV varies from the edge to the middle position, which induces a directional movement of surface-generated electrons, leading to effective photogenerated carrier separation compared with pristine BTO (Supplementary Fig. 20). Combined with the previous results, we find that the electron transfer is related to the spatial morphology and structure. The hollow regions with etched edges possess abundant photogenerated electrons, which is mainly because the regions with thin edges can accelerate electron migration to the surface from the bulk and center. In order to further determine the transfer and the regional distribution of electrons, the deposition of Au on BTO was performed with and without acid treatment. The random distribution of Au nanoparticle can be observed on the surfaces of pristine BTO in Fig. 3h. Not surprisingly, it is found to be preferentially deposited on the edge area of surface crystal facets of BTO-2 in Fig. 3i, which shows a special behavior, that is, spatially differentiated structure induces the differentiated distribution of electrons. A fast electron transfer dynamic will be promoted due to the electron differentiated distribution, which is beneficial to photogenerated charge separation for efficient OWS (Fig. 3j). The process of photocatalysis is largely limited by the separation and transfer of photogenerated charge carriers. The lifetime and transfer dynamic were characterized by photoluminescence (PL) and time-resolved photoluminescence (TRPL). Both pristine BTO and etched BTO show an evident PL peak at 485 nm (Supplementary Fig. 21). The PL intensity of the etched BTO gradually decreases with etching time, indicating a decrease in the number of excitons due to carries’ effective separation. We carried out TRPL to monitor the charge carrier dynamics on the nanosecond time-scale. As shown in the Fig. 4a, the carrier lifetime curve conforms to a multi-exponential decay function, which can be fitted by different parameters. The average lifetime is 3.64 ns for pristine BTO, which is longer than that of the etched BTO. And the average carrier lifetime gradually reduces with the duration of HCl treatment (3.28 ns for BTO-0.25, 2.71 ns for BTO-0.5, 2.34 ns for BTO-1, and 1.69 ns for BTO-2), representing that charge carriers’ transfer and lifetime is dependent on spatial morphology. Specifically, the faster decay lifetime of τ1 of etched BTO is longer than that of pristine BTO and the contribution of fast decay of τ1 increases largely from 34% for BTO to 85% for BTO-2, confirming the suppression of the inter-band recombination and fast electron transfer of the etched BTO due to the homojunction between the edge and the center. Also, combined with the morphological analysis, the decreased lifetime is gentle at the initial stages of etching, which is because the etching time is relatively short and only the surface Bi–O layer is etched (Fig. 4b). With increasing duration of the etching, the carrier’s lifetime of the sample rapidly declines, which is positively correlated to the etching depth due to the formation of the homojunction and the ultrathin edge, where the electrons of the etched edge will be accelerated and transferred to the surface compared with the center part of the BTO (Fig. 4c). Moreover, we further demonstrated the rapid interfacial electron transport performance. The etched BTO shows a larger photocurrent than that of the pristine BTO. And it is found that a positively correlation of the photocurrent with the etching depth for etched BTO (Supplementary Fig. 22a). The electrochemical impedance spectroscopy (EIS) measurements and analyses also reveal fast electron transport of the etched BTO. The radius of the fitted curve reflects the strength of the transfer of semiconductor electrons, that is, the smaller the radius indicates the faster the electron transfer. As shown in the impedance curves (Supplementary Fig. 22b), the BTO-2 shows the smallest electrochemical impedance, implying the fastest electron transport among the photocatalysts.
Synthesis and characterization of BTO photocatalysts As a layered structural two-dimensional material, BTO is characterized by regular stacking of bismuth oxygen [Bi2O2]2+ slabs and perovskite-like [Bi2Ti3O10]2– blocks, in which Bi–O bond can be etched via the interaction of proton and halide ion. Briefly, pristine single-crystal BTO was firstly prepared by a salt-assisted solid-state reaction. Due to the different structural stability of exposed crystal facets, the [Bi2O2]2+ structure of {001} crystal facet and Bi–O bond of {010} crystal facet of BTO can be selectively preferentially etched (Supplementary Fig. 1 and Supplementary Note 1). Also, the Ti–O bond of BTO {001} will prevent further etching of {001} crystal facet after etching [Bi2O2]2+, ultimately leading to the formation of a hollow structure along the edge region. Scanning electron microscopy (SEM) images show the pristine BTO with an evident nanosheet morphology. When HCl etching was conducted, the etched edge gradually became highly visible with the extension of the etching time to 2 h (Fig. 1a and Supplementary Fig. 2). However, the particle size or crystallite size of samples had no significant change, but the obvious difference of specific surface area occurs before and after acid etching (Supplementary Fig. 3 and Supplementary Table 1). When the etching time extended to 4 h, the BTO-4 shows that the nanosheets are broken with impurities due to nucleation of TiO2 (Supplementary Fig. 4). X-ray diffraction (XRD) pattern indicates the same crystal structure between the pristine and the etched BTO within 2 h (Fig. 1b). However, the crystallinity of the etched BTO is gradually weakened with further extension of etching time (Fig. 1c). Unsurprisingly, the over-etched BTO-4 shows a mixed structure of BTO and TiO2 (Supplementary Figs. 5 to 6 and Supplementary Note 2). Similarly, under the same concentration as HCl, the HBr-treated BTO sample shows similar crystal structure and morphology (Supplementary Figs. 7 and 8). The control experiments were performed with HNO3 and H2SO4 treated pristine BTO under the same concentration as HCl. The XRD results indicate the same crystal structure as BTO (Supplementary Fig. 9). And the edge region of BTO has not been etched (Supplementary Fig. 10). Further, an atomic force microscope (AFM) was used to investigate the thickness change before and after etching for different periods. No significant change appears in the overall thickness before and after etching and between the center and edge of the etched BTO (Supplementary Fig. 11). Furthermore, a highly hollow edge structure was observed by the SEM image, and the corresponding elemental mapping reveals the uniform distributions of Bi and Ti elements (Supplementary Fig. 12). Other SEM images, the corresponding elemental mapping, and linear scan elemental analysis are provided in Supplementary Figs. 13–15, and the results proved the uniform distributions of Bi, Ti, and O elements in both the center and the edge region of BTO with and without HCl treatment. Further characterization of the local structure of the etched BTO was carried out under high-resolution transmission electron microscopy (HRTEM). An evident thickness variation contrast occurred between the central and the edge regions of transmission electron microscopy (TEM) image because of the etching effect (Fig. 2a). The HRTEM image exhibits the same lattice orientation and lattice fringes between the edge and the central regions of the etched BTO (Figs. 2b,c), indicating the single crystalline nature of the structure. Furthermore, the same diffraction patterns further confirm the single-crystal structure by selected area electron diffraction (SAED) at different locations (Figs. 2d and 2e). Accordingly, the chemical valence states of elements were also characterized by X-ray photoelectron spectroscopy (XPS). Due to the gradually etched depth of the structure, the chemical states of the surface elements change in a dynamically evolving manner, especially for Ti and Bi. Evidently, as shown in Fig. 2f, the binding energy of Ti 2p gradually shifts towards a higher value, which comes from the increasing proportion of the Ti–O structures exposed on the surface because of the Ti element binding more high-energy O (hydroxyls and water molecules). Due to the change of lattice oxygen (Bi–O and Ti–O) structure on the exposed surface, the binding energy of O 1 s shifts to higher value at the same time (Supplementary Fig. 16a). In contrast, the corresponding Bi 4 f peak slightly shifts downward in binding energy with increasing extent of the acid etching, because the outside high binding energy Bi of [Bi2O2]2+ layer and [Bi2Ti3O10]2– was etched away. Although the binding energies of Bi and Ti evolve, the main structure of samples keeps unchanged during the etching process (Fig. 2g). No Cl element was observed during the etching process, indicating that the nucleation of bismuth oxyhalide was inhibited (Supplementary Fig. 16b). In addition, the structural change should be accompanied by the evolution of the band structure. The band gap slightly increases with the extension of the etching, which is due to the change of the band structure by the spatial size effect, which is mainly from the ultrathin hollow edge (Supplementary Figs. 17–19). With the increase of edge etching depth, the trend of band structure change is more obvious. Meanwhile, the difference in the band structures between the unetched part and adjacent etched BTO will form “homojunction”, which are beneficial to the transfer of electrons to the surface of the photocatalyst. Also, due to TiO2 impurities resulting from structural destruction caused by the excessive etching, the BTO-4 further shows the bigger band gap and evident differences from BTO in light absorption behavior. However, such a small change of bandgaps is not expected to influence the photocatalytic performance. Further understanding the charge distribution on the crystal surface is helpful to explore the process of photocatalytic water splitting. SPVM is an effective tool for spatial correlation between photogenerated charge distribution and structural morphology, as well as the dependence between light absorption and carrier migration and transfer46,47. As shown in Figs. 3a–f, photogenerated electrons were imaged in the surface due to negative SPV. With the increasing wavelength of illumination light, the intensity of the SPV decreases, which is consistent with the behavior of the light absorption spectrum, which proves that the charge generation is driven by light (Supplementary Fig. 17 and Fig. 3g). The difference is that when the light irradiated the surface of photocatalyst, the charge distribution intensity varies between the edge and the middle of the sample. This phenomenon is very interesting and differs from most of the results reported in the literatures. We speculate the carrier distribution is dependent on the space geometry. Differential SPV analysis was performed at different positions of a given facet and the differences were compared at the three positions. The corresponding results are shown in Figs. 3a–f. The SPV is significantly different with the change of light wavelength at the three locations of α1, α2 and α3. Therefore, the intensity of SPV varies from the edge to the middle position, which induces a directional movement of surface-generated electrons, leading to effective photogenerated carrier separation compared with pristine BTO (Supplementary Fig. 20). Combined with the previous results, we find that the electron transfer is related to the spatial morphology and structure. The hollow regions with etched edges possess abundant photogenerated electrons, which is mainly because the regions with thin edges can accelerate electron migration to the surface from the bulk and center. In order to further determine the transfer and the regional distribution of electrons, the deposition of Au on BTO was performed with and without acid treatment. The random distribution of Au nanoparticle can be observed on the surfaces of pristine BTO in Fig. 3h. Not surprisingly, it is found to be preferentially deposited on the edge area of surface crystal facets of BTO-2 in Fig. 3i, which shows a special behavior, that is, spatially differentiated structure induces the differentiated distribution of electrons. A fast electron transfer dynamic will be promoted due to the electron differentiated distribution, which is beneficial to photogenerated charge separation for efficient OWS (Fig. 3j). The process of photocatalysis is largely limited by the separation and transfer of photogenerated charge carriers. The lifetime and transfer dynamic were characterized by photoluminescence (PL) and time-resolved photoluminescence (TRPL). Both pristine BTO and etched BTO show an evident PL peak at 485 nm (Supplementary Fig. 21). The PL intensity of the etched BTO gradually decreases with etching time, indicating a decrease in the number of excitons due to carries’ effective separation. We carried out TRPL to monitor the charge carrier dynamics on the nanosecond time-scale. As shown in the Fig. 4a, the carrier lifetime curve conforms to a multi-exponential decay function, which can be fitted by different parameters. The average lifetime is 3.64 ns for pristine BTO, which is longer than that of the etched BTO. And the average carrier lifetime gradually reduces with the duration of HCl treatment (3.28 ns for BTO-0.25, 2.71 ns for BTO-0.5, 2.34 ns for BTO-1, and 1.69 ns for BTO-2), representing that charge carriers’ transfer and lifetime is dependent on spatial morphology. Specifically, the faster decay lifetime of τ1 of etched BTO is longer than that of pristine BTO and the contribution of fast decay of τ1 increases largely from 34% for BTO to 85% for BTO-2, confirming the suppression of the inter-band recombination and fast electron transfer of the etched BTO due to the homojunction between the edge and the center. Also, combined with the morphological analysis, the decreased lifetime is gentle at the initial stages of etching, which is because the etching time is relatively short and only the surface Bi–O layer is etched (Fig. 4b). With increasing duration of the etching, the carrier’s lifetime of the sample rapidly declines, which is positively correlated to the etching depth due to the formation of the homojunction and the ultrathin edge, where the electrons of the etched edge will be accelerated and transferred to the surface compared with the center part of the BTO (Fig. 4c). Moreover, we further demonstrated the rapid interfacial electron transport performance. The etched BTO shows a larger photocurrent than that of the pristine BTO. And it is found that a positively correlation of the photocurrent with the etching depth for etched BTO (Supplementary Fig. 22a). The electrochemical impedance spectroscopy (EIS) measurements and analyses also reveal fast electron transport of the etched BTO. The radius of the fitted curve reflects the strength of the transfer of semiconductor electrons, that is, the smaller the radius indicates the faster the electron transfer. As shown in the impedance curves (Supplementary Fig. 22b), the BTO-2 shows the smallest electrochemical impedance, implying the fastest electron transport among the photocatalysts.
考点1:“Homojunction” 必须译为 ”同质结” 考点2:“Salt-assisted solid-state reaction” 推荐译为 ”熔盐辅助固相反应” 考点3:“Over-etched” 必须译为 ”过度刻蚀” 考点4:“Carrier migration and transfer” 必须译为”载流子迁移与传输” 考点5:“Electron differentiated distribution” 必须译为 ”电子差异化分布” 考点6:“Particle size” 推荐译为 ”粒径/颗粒尺寸” 考点7:“Control experiments” 必须译为 ”对照实验” 考点8:“Dynamically evolving manner” 必须译为 ”动态演化方式”
5bcc3
学术论文
自然科学
144
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Summary Why did some (but not all) Chinese dynasties invest huge amounts of resources in the construction of ‘Great Walls’? The proposed project will focus on precisely that question, in an attempt to unravel what is, perhaps, the most enigmatic episode of ‘Great Wall’ construction. Roughly dated to the 10th-13th centuries CE and located far to the north of other ‘Great Wall’ lines, this Medieval Wall System (MWS) is one of the longest walls ever constructed in world history, stretching over more than 3,500 km and including large auxiliary structures. The amount of resources invested in this MWS must have been enormous, but historical sources are mute about its construction, and modern scholarship is unable to date it precisely or understand why it was built and how it functioned. The motives behind the construction of the MWS, its political context and ecological implications, are highly relevant for the understanding of the complex history of China and Mongolia on the eve of Chinggis Khan’s rise to power. However, because in the past scholars have assumed that ‘Great Walls’ were fortified border lines designed to stop military incursions, such issues’ impetus and consequences were never addressed. Hence, the proposed project will put forward novel hypotheses, analyse them by using advanced recovery and analytical methods, and examine them against a broad archaeological, historical, environmental, and geographical background. The research hypothesis of the proposed project is that the MWS was not built as a defence against invading armies, but rather as a means to monitor and sometimes stop the movement of nomadic people and their herds. The large scale movements of nomadic people towards more central areas of the empire happened, I would suggest, in times of ecological stress in the Steppe. Objectives Why did some (but not all) Chinese dynasties invest huge amounts of resources in the construction of ‘Great Walls’? The proposed project will focus on precisely that question, in an attempt to unravel what is, perhaps, the most enigmatic episode of ‘Great Wall’ construction. This wall system is located in North China and Mongolia and covers a distance of over 3,500 km. However, if one takes into account the different parallel lines, their accumulated length is probably closer to 6,500 km, making it one of the longest walls ever constructed in world history. Notwithstanding its huge size and the vast resources, material and human, that must have been invested in its construction, it is unclear who built it and for what purpose(s). The construction of this complex system, which includes long earthen walls and accompanying ditches, auxiliary structures (camps, beacon towers, etc.) and roads, is dated roughly to the 10th to 13th centuries CE, but its construction is not described in any of the relevant historical chronicles. Moreover, very little historical and archaeological research has been done on it. This Medieval Wall System (MWS) is enigmatic not only because its history is obscure, but also because it is located deep in the Steppe area, hundreds of kilometres north and west of earlier and later lines of the so-called ‘Great Wall of China’. The proposed project shall address some of the fundamental questions regarding the MWS: When was it built and for how long was it in use? Who built it? For what purposes? What was the structure of the MWS (including the associated features) and how did it function? What amount of human labour and other resources were necessary in order to build it? On a more fundamental level, this project will attempt to answer the more basic question: What was the logic behind the construction of such long and enormously expensive linear barriers (‘Great Walls’) by several Chinese dynasties (and other polities in the pre-modern world)? What were the political preconditions for such monumental works? How, if at all, did wall construction relate to climatic changes and ecological conditions? These issues were not addressed properly in the past, first and foremost because scholars have assumed that ‘Great Walls’ were fortified border lines designed to stop military incursions. The proposed research will suggest novel hypotheses, never previously considered, and examine them using advanced recovery and analytical methods, integrating archaeological, historical, environmental, and geographical data. Research Hypothesis and Expected Breakthroughs The construction of long-walls is not a Chinese phenomenon, as they are also known from other ancient civilizations throughout the Old World. Notable examples are the Roman limes and the Sasanian Gorgān and Tammīsheh walls. However, the most extensive examples in time, space and amount of labour invested in their construction, are the walls built by Chinese empires. The so-called ‘Great Wall of China’ is, in fact, a series of walls constructed by different dynasties at different times. Each of those walls extends over hundreds, sometimes even thousands, of kilometres and is accompanied by auxiliary structures (camps, beacon towers, etc.) and roads. Estimates of the amount of work required to build these walls vary, but it was suggested, for example, that it took a working force of at least 300,000 men five years to construct the earliest, and by no means the most extensive, Chinese wall system. While Chinese dynasties constructed the most extensive wall-systems the world has ever seen, many dynasties in China did not construct long-walls. Apart from the construction of several different walls by pre-dynastic states of the Warring States period (481-221 BCE), long-walls were constructed during four main episodes over the last two millennia: the Qin and Western Han dynasties (the 3rd to 1st centuries BCE); the Northern Wei and Northern Qi dynasties (5th and 6th centuries CE); the Liao-Jin dynasties (11th or 12th and 13th centuries); and the Ming dynasty (15th and 16th centuries). A traditional view describes the ‘Great Wall’ as a symbolic and real demarcation of a cultural-ecological line that separates two distinct ways of life: the pastoral-nomadism of the Steppe and desert areas, on the one hand, and sedentary agricultural societies, on the other. Consequently, modern scholarship identifies the walls as a fortified border intended to stop nomadic armies from invading China. This paradigm of Chinese long-walls (including the MWS) as military installations has never been challenged, although disagreement regarding the motivation for ‘Great Wall’ construction exists between scholars who see them as defensive structures against the aggression of nomadic tribes and those who argue that ‘Great Walls’ were part of the Chinese imperial expansion and conquest of lands previously inhabited by nomads. And yet, a superficial examination of the MWS suggests that it is at odds with many of the fundamental assumptions of the accepted paradigm. For example, the MWS is located deep in the Steppe zone (and, in a few of its southern parts, inside the desert area), and does not demarcate any cultural-ecological border. Moreover, although we do not know who constructed this wall, the two main candidates are the Liao and the Jin dynasties, both of which were founded by nomadic or semi-nomadic peoples, and subsequently their concept of empire had more to do with control over people, many of them nomads, rather than forming clearly defined land borders. In addition, a preliminary analysis based on pilot field work and GIS analysis suggests that the location and structure of the wall and the auxiliary structures associated with it are not compatible with what we might expect from a military installation that was meant to defend the border against invading armies. Based on this preliminary information, the research hypothesis of the proposed project is that the MWS, and probably many of the other episodes of ‘Great Wall’ construction in Chinese history, were built to monitor and sometimes stop movements of people rather than armies. It is highly plausible that the walls were built during periods of large-scale movements of nomadic people towards more central areas of the empire. An additional hypothesis, as to why such control was needed, suggests that extreme climatic anomalies, especially concentrations of cold spells and droughts, must have affected the fragile ecology of the Steppe. The devastating effects of these climatic phenomena, known as dzud in Mongolian, on communities of nomadic pastoralists, are well documented in historical and modern periods. For example, between 2000-2002, and again in the winter of 2009-2010, intense cold conditions caused the death of 20% to 40% of all livestock, causing mass migration of herders to the Mongolian capital, Ulaanbaatar. Whereas sporadic and local occurrences of such events are common, a concentration of extreme events over a period of ten to twenty years must have an acute socio-economic impact. The 10th-14th centuries were a period of climatic instability, known as the Medieval Climate Anomaly. Although this global phenomenon did not have the same effect on the local ecologies of different places throughout Eurasia, recent research has started to flesh out the ways in which it catalysed economic crisis, socio-political unrest, and large-scale migration in the Mediterranean and West Asia, as well as in East Asia. If these hypotheses (or some of them) are confirmed for the MWS, it will change our understanding of the complex history of the region during the centuries prior to Chinggis Khan’s rise to power. These hypotheses may not account for all the episodes of ‘Great Wall’ construction in China. For example, the wall constructed by the Ming dynasty is much more massive than that of earlier periods. Although it was not successful in defending the Ming, there are historic accounts of it being a formidable military obstacle. However, the model developed by this research can serve to explain some of the other episodes of ‘Great Wall’ construction in Chinese history and change our perception of the conditions that triggered wall building in China, the purposes for which these walls were built, and their function. Understanding the background and reasons for wall construction will thus contribute to a better understanding of the decision and policy-making of Chinese dynasties, and of the dynastic system more generally. Such insights could be applicable to the analysis of long-walls of other civilizations. For example, the famous Hadrian’s Wall, which is seen by many as the fortified border line par excellence, has a gate every Roman mile (1,479 m). It is suggested that creating so many openings in the wall was intended to enable, but also to supervise, the civilian movement of traders and, perhaps, transhumance.
Summary Why did some (but not all) Chinese dynasties invest huge amounts of resources in the construction of ‘Great Walls’? The proposed project will focus on precisely that question, in an attempt to unravel what is, perhaps, the most enigmatic episode of ‘Great Wall’ construction. Roughly dated to the 10th-13th centuries CE and located far to the north of other ‘Great Wall’ lines, this Medieval Wall System (MWS) is one of the longest walls ever constructed in world history, stretching over more than 3,500 km and including large auxiliary structures. The amount of resources invested in this MWS must have been enormous, but historical sources are mute about its construction, and modern scholarship is unable to date it precisely or understand why it was built and how it functioned. The motives behind the construction of the MWS, its political context and ecological implications, are highly relevant for the understanding of the complex history of China and Mongolia on the eve of Chinggis Khan’s rise to power. However, because in the past scholars have assumed that ‘Great Walls’ were fortified border lines designed to stop military incursions, such issues’ impetus and consequences were never addressed. Hence, the proposed project will put forward novel hypotheses, analyse them by using advanced recovery and analytical methods, and examine them against a broad archaeological, historical, environmental, and geographical background. The research hypothesis of the proposed project is that the MWS was not built as a defence against invading armies, but rather as a means to monitor and sometimes stop the movement of nomadic people and their herds. The large scale movements of nomadic people towards more central areas of the empire happened, I would suggest, in times of ecological stress in the Steppe. Objectives Why did some (but not all) Chinese dynasties invest huge amounts of resources in the construction of ‘Great Walls’? The proposed project will focus on precisely that question, in an attempt to unravel what is, perhaps, the most enigmatic episode of ‘Great Wall’ construction. This wall system is located in North China and Mongolia and covers a distance of over 3,500 km. However, if one takes into account the different parallel lines, their accumulated length is probably closer to 6,500 km, making it one of the longest walls ever constructed in world history. Notwithstanding its huge size and the vast resources, material and human, that must have been invested in its construction, it is unclear who built it and for what purpose(s). The construction of this complex system, which includes long earthen walls and accompanying ditches, auxiliary structures (camps, beacon towers, etc.) and roads, is dated roughly to the 10th to 13th centuries CE, but its construction is not described in any of the relevant historical chronicles. Moreover, very little historical and archaeological research has been done on it. This Medieval Wall System (MWS) is enigmatic not only because its history is obscure, but also because it is located deep in the Steppe area, hundreds of kilometres north and west of earlier and later lines of the so-called ‘Great Wall of China’. The proposed project shall address some of the fundamental questions regarding the MWS: When was it built and for how long was it in use? Who built it? For what purposes? What was the structure of the MWS (including the associated features) and how did it function? What amount of human labour and other resources were necessary in order to build it? On a more fundamental level, this project will attempt to answer the more basic question: What was the logic behind the construction of such long and enormously expensive linear barriers (‘Great Walls’) by several Chinese dynasties (and other polities in the pre-modern world)? What were the political preconditions for such monumental works? How, if at all, did wall construction relate to climatic changes and ecological conditions? These issues were not addressed properly in the past, first and foremost because scholars have assumed that ‘Great Walls’ were fortified border lines designed to stop military incursions. The proposed research will suggest novel hypotheses, never previously considered, and examine them using advanced recovery and analytical methods, integrating archaeological, historical, environmental, and geographical data. Research Hypothesis and Expected Breakthroughs The construction of long-walls is not a Chinese phenomenon, as they are also known from other ancient civilizations throughout the Old World. Notable examples are the Roman limes and the Sasanian Gorgān and Tammīsheh walls. However, the most extensive examples in time, space and amount of labour invested in their construction, are the walls built by Chinese empires. The so-called ‘Great Wall of China’ is, in fact, a series of walls constructed by different dynasties at different times. Each of those walls extends over hundreds, sometimes even thousands, of kilometres and is accompanied by auxiliary structures (camps, beacon towers, etc.) and roads. Estimates of the amount of work required to build these walls vary, but it was suggested, for example, that it took a working force of at least 300,000 men five years to construct the earliest, and by no means the most extensive, Chinese wall system. While Chinese dynasties constructed the most extensive wall-systems the world has ever seen, many dynasties in China did not construct long-walls. Apart from the construction of several different walls by pre-dynastic states of the Warring States period (481-221 BCE), long-walls were constructed during four main episodes over the last two millennia: the Qin and Western Han dynasties (the 3rd to 1st centuries BCE); the Northern Wei and Northern Qi dynasties (5th and 6th centuries CE); the Liao-Jin dynasties (11th or 12th and 13th centuries); and the Ming dynasty (15th and 16th centuries). A traditional view describes the ‘Great Wall’ as a symbolic and real demarcation of a cultural-ecological line that separates two distinct ways of life: the pastoral-nomadism of the Steppe and desert areas, on the one hand, and sedentary agricultural societies, on the other. Consequently, modern scholarship identifies the walls as a fortified border intended to stop nomadic armies from invading China. This paradigm of Chinese long-walls (including the MWS) as military installations has never been challenged, although disagreement regarding the motivation for ‘Great Wall’ construction exists between scholars who see them as defensive structures against the aggression of nomadic tribes and those who argue that ‘Great Walls’ were part of the Chinese imperial expansion and conquest of lands previously inhabited by nomads. And yet, a superficial examination of the MWS suggests that it is at odds with many of the fundamental assumptions of the accepted paradigm. For example, the MWS is located deep in the Steppe zone (and, in a few of its southern parts, inside the desert area), and does not demarcate any cultural-ecological border. Moreover, although we do not know who constructed this wall, the two main candidates are the Liao and the Jin dynasties, both of which were founded by nomadic or semi-nomadic peoples, and subsequently their concept of empire had more to do with control over people, many of them nomads, rather than forming clearly defined land borders. In addition, a preliminary analysis based on pilot field work and GIS analysis suggests that the location and structure of the wall and the auxiliary structures associated with it are not compatible with what we might expect from a military installation that was meant to defend the border against invading armies. Based on this preliminary information, the research hypothesis of the proposed project is that the MWS, and probably many of the other episodes of ‘Great Wall’ construction in Chinese history, were built to monitor and sometimes stop movements of people rather than armies. It is highly plausible that the walls were built during periods of large-scale movements of nomadic people towards more central areas of the empire. An additional hypothesis, as to why such control was needed, suggests that extreme climatic anomalies, especially concentrations of cold spells and droughts, must have affected the fragile ecology of the Steppe. The devastating effects of these climatic phenomena, known as dzud in Mongolian, on communities of nomadic pastoralists, are well documented in historical and modern periods. For example, between 2000-2002, and again in the winter of 2009-2010, intense cold conditions caused the death of 20% to 40% of all livestock, causing mass migration of herders to the Mongolian capital, Ulaanbaatar. Whereas sporadic and local occurrences of such events are common, a concentration of extreme events over a period of ten to twenty years must have an acute socio-economic impact. The 10th-14th centuries were a period of climatic instability, known as the Medieval Climate Anomaly. Although this global phenomenon did not have the same effect on the local ecologies of different places throughout Eurasia, recent research has started to flesh out the ways in which it catalysed economic crisis, socio-political unrest, and large-scale migration in the Mediterranean and West Asia, as well as in East Asia. If these hypotheses (or some of them) are confirmed for the MWS, it will change our understanding of the complex history of the region during the centuries prior to Chinggis Khan’s rise to power. These hypotheses may not account for all the episodes of ‘Great Wall’ construction in China. For example, the wall constructed by the Ming dynasty is much more massive than that of earlier periods. Although it was not successful in defending the Ming, there are historic accounts of it being a formidable military obstacle. However, the model developed by this research can serve to explain some of the other episodes of ‘Great Wall’ construction in Chinese history and change our perception of the conditions that triggered wall building in China, the purposes for which these walls were built, and their function. Understanding the background and reasons for wall construction will thus contribute to a better understanding of the decision and policy-making of Chinese dynasties, and of the dynastic system more generally. Such insights could be applicable to the analysis of long-walls of other civilizations. For example, the famous Hadrian’s Wall, which is seen by many as the fortified border line par excellence, has a gate every Roman mile (1,479 m). It is suggested that creating so many openings in the wall was intended to enable, but also to supervise, the civilian movement of traders and, perhaps, transhumance.
考点1:“Roman limes”必须译为“罗马帝国界墙”或者“罗马帝国长城”,因为这是固定翻译,译文选取一种即可,保持译文一致性 考点2:“Gorgān and Tammīsheh walls”必须译为“戈尔干长城和塔米舍长城”,固定景点翻译 考点3:“pre-dynastic states”推荐译为“封建王朝形成之前的政权”,不能简单翻译成“前王朝时期”,会容易被非专业读者误解称为古埃及王朝形成之前的时期 考点4:“dzud ” 推荐译为“严冬灾害期” 考点5:“Roman mile”必须译为“罗里”,计数单位的固定译法
5ef50
学术论文
人文科学
194
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: A client of mine, Melissa, the CEO of a tech company in the healthcare sector, vented to me about a conversation she’d had with her CHRO, Ben. He came to her and said, “I’m hearing frustrations on the team about how decisions are getting made.” Having prided herself on being a transparent, inclusive leader (which she is), she was perplexed. Ben went on to share several stories of team members who’d learned about critical decisions that affected them from their peers, not from Melissa. For example, the VP of engineering learned about a deprioritized product feature from the VP of marketing. The VP of quality learned about a new project to accelerate product delivery from the VP of operations. The VP of sales learned about a new budget shift from the CFO. Melissa explained to me that, indeed, based on new information that each of those executives had brought in their 1:1s, she supported their recommended decisions, and assumed they would loop in the requisite peers rather than waiting for the next monthly executive meeting. In her mind, she was empowering her team to accelerate needed decisions rather than being the bottleneck or playing messenger to the other VPs, thus slowing things down. Clearly her focus on efficiency blinded her to the disunity and conflict it might provoke on the team. In each case, she’d only gotten part of the story. I gave her the advice I’ve given countless CEOs and C-suite leaders: Stop having so many 1:1 meetings. The Hidden Costs of Executive 1:1 Meetings In most large organizations, a typical CEO’s or senior executive’s calendar is clogged with 1:1 meetings. These are usually seen as necessary for alignment, decision-making, or relationship management. But at the top of an enterprise, the very structure of these meetings is working against the organization’s best interests. While there’s lots of information available about how to optimize or improve your 1:1 meetings, no amount of improvement will help meetings that shouldn’t be happening in the first place. Lower in the organization, 1:1 meetings serve different purposes that may make more sense when people are guiding middle managers, supervisors, or individual contributors. But here’s the counterintuitive truth: The more 1:1 meetings senior executives have, the more fragmented and functionally siloed the organization can become. It might be easy to conclude that Melissa should have known better not to assume her functional VPs had consulted with key peers impacted by their decisions, or that she simply should have said, “Before we move ahead, why don’t you loop John in to make sure this doesn’t cause any issues for him.” Fair enough. But even if that had happened, it wouldn’t entirely eliminate the challenges of fragmentation caused by executive 1:1s. And worse, it may have led to five additional meetings she would have had to referee. At the executive level, no meaningful decision exists in a vacuum. As was the case with Melissa’s team, what the engineering VP does is significantly impacted by a decision the VP of marketing makes. The VP of operations’ decision had significant implications for the VP of quality. Yet Melissa’s meetings with those key leaders resulted in decisions—and assumptions—that others didn’t hear, couldn’t weigh in on, or later had to be retroactively briefed about. There was no intended malice or attempt to usurp others’ influence. Just the natural centrifugal force of a myopic view of the world through one’s own function or region. This creates four core problems: Fragmented governance Each 1:1 functions like a mini steering committee, but with no one else in the room. Governance becomes informal and duplicative, requiring rework and follow-up meetings just to keep others up to speed. Worse, viewpoints that weren’t considered in the initial 1:1 then surface later on, leading to friction that could have been avoided. When too much time is spent in fragmented meetings, there’s little left for the work that only senior leaders can do: stewarding enterprise priorities, developing talent, and advancing culture. Functional bias One-on-ones reinforce a view of the organization as a set of departments rather than a web of capabilities. Over time, this strengthens silos instead of creating cross-functional coherence. And it forces the CEO or senior leader to act as the sole integrator of cross-functional perspectives. Decision repackaging Executives waste valuable time repeating, translating, or defending decisions made privately. This not only slows momentum—it breeds misalignment and friction downstream. Trust erodes over time, and leaders become conditioned to use their 1:1s to game the system and distort the truth to ensure their views prevail. Executive rivalry and collusion Succession overhang and ambition can subtly foster competition among top leaders. One-on-ones become arenas for inside influence, where leaders can imply advantage by referencing private access (“In my 1:1 with Bill this morning…”), creating a sense of secrecy and status that erodes collective trust. CEOs can also be guilty of leveraging their 1:1s to extract information about other team members, creating a false sense of “insider intimacy” with direct reports that ultimately undermines psychological safety. One CEO I urged to reduce his 1:1s actually lamented to me, “But how will I get all of the inside scoops I don’t get anywhere else?” A Better Way to Use Executives’ Time: Capability Meetings The alternative is to re-engineer the way executive time is used. Instead of relying on 1:1s for operational discussions, CEOs and senior executives should convene small, cross-functional “capability meetings”—1:2 or 1:3 conversations that reflect how value is actually created. For example, innovation isn’t owned by R&D, marketing, or customer experience alone; it lives in the seams between them. Bringing those leaders together—at the same table—ensures strategic clarity, accelerates coordination, and strengthens shared ownership. Here are five ways to begin: Reserve 1:1s for development. Shift 1:1s to a quarterly cadence and use them solely for the individual’s growth, feedback, and career development. These conversations deserve focused, undivided attention—not to be diluted by operational updates or real-time firefighting. You can block 90-minute quarterly sessions that are clearly labeled “development conversation.” Structure them around reflection, feedback, and future goals. Avoid sliding into tactical issues. Use prompts like “What are you learning about yourself as a leader this quarter?” or “What would stretch you in a meaningful way next year?” At one global consumer goods company, the COO redesigned her 1:1 structure by implementing a “career check-in” model: quarterly sessions focused entirely on each direct report’s long-term goals, with zero slides or project updates. This shift dramatically improved retention and succession visibility within her team. Convene around capabilities. Map out your enterprise’s most critical capabilities like innovation, digital transformation, or customer loyalty, and structure standing meetings around the cross-functional intersections that enable them. Identify the five to seven capabilities that create differentiated value in your business. For each, bring together the two to four functions most essential to delivering that value. These “capability councils” become the forums for near-term decision-making, trade-offs, and shared execution, saving the full executive team’s time for truly enterprise-wide strategies and opportunities. Melissa worked with her team to reframe their strategic priorities into capability areas—like “virtual care experience” and “clinical operations technologies”—and created standing 1:3 capability meetings that included product, operations, and engineering leaders. As a result, go-to-market cycles shortened by 20%, and duplication and friction across teams dropped significantly. Ensure the right people are listening. One of the quiet inefficiencies of 1:1s is that they pull decisions into private channels, where the people most affected often aren’t present. That means leaders have to replay discussions for others later, or relitigate decisions once new information surfaces, leading to misunderstanding, misalignment, or delayed action. You can use capability meetings to bring together the specific leaders whose functions intersect around a given issue. This ensures that the people who need to understand and act on the information are hearing it at the moment it matters. No one has to wait to be briefed later. This kind of targeted presence builds sharper shared judgment and eliminates the need to re-translate decisions across silos. At a financial services firm I’ve consulted with, the CEO replaced many of his 1:1s with capability meetings like “go to market,” “operational efficiency,” or “enterprise talent,” convening various combinations of the heads of risk, compliance, sales, and human resources, as well as the business unit general managers, to discuss capability-focused topics. Because the right people were hearing the context in real time, execution accelerated. Leaders no longer had to “sell” decisions afterward or decipher what had happened behind closed doors because they were in the room when it happened. Elevate executive team time. When too many decisions happen in private 1:1s, full executive team meetings often devolve into catch-up sessions, replaying decisions, clarifying miscommunications, or managing interpersonal fallout. That robs the team of the chance to focus on the truly enterprise-wide work only they can do. By using capability meetings to handle cross-functional execution and trade-offs at the right level, executive team time can be reserved for strategic sensemaking, cultural stewardship, and long-horizon bets. Center agendas around questions like “What are the few issues that require the full weight of this team?” or “Where does our system need realignment?” One Fortune 100 company overhauled its monthly executive team meeting after realizing that too much time was spent relitigating decisions made elsewhere. By reassigning operational decision-making to triads tied to core capabilities, they cleared the decks for their senior team to focus on systemic enterprise priorities—like cultural resilience, the innovation pipeline, scenario planning for disruptions, and global portfolio shifts. The result was faster execution at the edges and deeper strategic coherence at the center.
A client of mine, Melissa, the CEO of a tech company in the healthcare sector, vented to me about a conversation she’d had with her CHRO, Ben. He came to her and said, “I’m hearing frustrations on the team about how decisions are getting made.” Having prided herself on being a transparent, inclusive leader (which she is), she was perplexed. Ben went on to share several stories of team members who’d learned about critical decisions that affected them from their peers, not from Melissa. For example, the VP of engineering learned about a deprioritized product feature from the VP of marketing. The VP of quality learned about a new project to accelerate product delivery from the VP of operations. The VP of sales learned about a new budget shift from the CFO. Melissa explained to me that, indeed, based on new information that each of those executives had brought in their 1:1s, she supported their recommended decisions, and assumed they would loop in the requisite peers rather than waiting for the next monthly executive meeting. In her mind, she was empowering her team to accelerate needed decisions rather than being the bottleneck or playing messenger to the other VPs, thus slowing things down. Clearly her focus on efficiency blinded her to the disunity and conflict it might provoke on the team. In each case, she’d only gotten part of the story. I gave her the advice I’ve given countless CEOs and C-suite leaders: Stop having so many 1:1 meetings. The Hidden Costs of Executive 1:1 Meetings In most large organizations, a typical CEO’s or senior executive’s calendar is clogged with 1:1 meetings. These are usually seen as necessary for alignment, decision-making, or relationship management. But at the top of an enterprise, the very structure of these meetings is working against the organization’s best interests. While there’s lots of information available about how to optimize or improve your 1:1 meetings, no amount of improvement will help meetings that shouldn’t be happening in the first place. Lower in the organization, 1:1 meetings serve different purposes that may make more sense when people are guiding middle managers, supervisors, or individual contributors. But here’s the counterintuitive truth: The more 1:1 meetings senior executives have, the more fragmented and functionally siloed the organization can become. It might be easy to conclude that Melissa should have known better not to assume her functional VPs had consulted with key peers impacted by their decisions, or that she simply should have said, “Before we move ahead, why don’t you loop John in to make sure this doesn’t cause any issues for him.” Fair enough. But even if that had happened, it wouldn’t entirely eliminate the challenges of fragmentation caused by executive 1:1s. And worse, it may have led to five additional meetings she would have had to referee. At the executive level, no meaningful decision exists in a vacuum. As was the case with Melissa’s team, what the engineering VP does is significantly impacted by a decision the VP of marketing makes. The VP of operations’ decision had significant implications for the VP of quality. Yet Melissa’s meetings with those key leaders resulted in decisions—and assumptions—that others didn’t hear, couldn’t weigh in on, or later had to be retroactively briefed about. There was no intended malice or attempt to usurp others’ influence. Just the natural centrifugal force of a myopic view of the world through one’s own function or region. This creates four core problems: Fragmented governance Each 1:1 functions like a mini steering committee, but with no one else in the room. Governance becomes informal and duplicative, requiring rework and follow-up meetings just to keep others up to speed. Worse, viewpoints that weren’t considered in the initial 1:1 then surface later on, leading to friction that could have been avoided. When too much time is spent in fragmented meetings, there’s little left for the work that only senior leaders can do: stewarding enterprise priorities, developing talent, and advancing culture. Functional bias One-on-ones reinforce a view of the organization as a set of departments rather than a web of capabilities. Over time, this strengthens silos instead of creating cross-functional coherence. And it forces the CEO or senior leader to act as the sole integrator of cross-functional perspectives. Decision repackaging Executives waste valuable time repeating, translating, or defending decisions made privately. This not only slows momentum—it breeds misalignment and friction downstream. Trust erodes over time, and leaders become conditioned to use their 1:1s to game the system and distort the truth to ensure their views prevail. Executive rivalry and collusion Succession overhang and ambition can subtly foster competition among top leaders. One-on-ones become arenas for inside influence, where leaders can imply advantage by referencing private access (“In my 1:1 with Bill this morning…”), creating a sense of secrecy and status that erodes collective trust. CEOs can also be guilty of leveraging their 1:1s to extract information about other team members, creating a false sense of “insider intimacy” with direct reports that ultimately undermines psychological safety. One CEO I urged to reduce his 1:1s actually lamented to me, “But how will I get all of the inside scoops I don’t get anywhere else?” A Better Way to Use Executives’ Time: Capability Meetings The alternative is to re-engineer the way executive time is used. Instead of relying on 1:1s for operational discussions, CEOs and senior executives should convene small, cross-functional “capability meetings”—1:2 or 1:3 conversations that reflect how value is actually created. For example, innovation isn’t owned by R&D, marketing, or customer experience alone; it lives in the seams between them. Bringing those leaders together—at the same table—ensures strategic clarity, accelerates coordination, and strengthens shared ownership. Here are five ways to begin: Reserve 1:1s for development. Shift 1:1s to a quarterly cadence and use them solely for the individual’s growth, feedback, and career development. These conversations deserve focused, undivided attention—not to be diluted by operational updates or real-time firefighting. You can block 90-minute quarterly sessions that are clearly labeled “development conversation.” Structure them around reflection, feedback, and future goals. Avoid sliding into tactical issues. Use prompts like “What are you learning about yourself as a leader this quarter?” or “What would stretch you in a meaningful way next year?” At one global consumer goods company, the COO redesigned her 1:1 structure by implementing a “career check-in” model: quarterly sessions focused entirely on each direct report’s long-term goals, with zero slides or project updates. This shift dramatically improved retention and succession visibility within her team. Convene around capabilities. Map out your enterprise’s most critical capabilities like innovation, digital transformation, or customer loyalty, and structure standing meetings around the cross-functional intersections that enable them. Identify the five to seven capabilities that create differentiated value in your business. For each, bring together the two to four functions most essential to delivering that value. These “capability councils” become the forums for near-term decision-making, trade-offs, and shared execution, saving the full executive team’s time for truly enterprise-wide strategies and opportunities. Melissa worked with her team to reframe their strategic priorities into capability areas—like “virtual care experience” and “clinical operations technologies”—and created standing 1:3 capability meetings that included product, operations, and engineering leaders. As a result, go-to-market cycles shortened by 20%, and duplication and friction across teams dropped significantly. Ensure the right people are listening. One of the quiet inefficiencies of 1:1s is that they pull decisions into private channels, where the people most affected often aren’t present. That means leaders have to replay discussions for others later, or relitigate decisions once new information surfaces, leading to misunderstanding, misalignment, or delayed action. You can use capability meetings to bring together the specific leaders whose functions intersect around a given issue. This ensures that the people who need to understand and act on the information are hearing it at the moment it matters. No one has to wait to be briefed later. This kind of targeted presence builds sharper shared judgment and eliminates the need to re-translate decisions across silos. At a financial services firm I’ve consulted with, the CEO replaced many of his 1:1s with capability meetings like “go to market,” “operational efficiency,” or “enterprise talent,” convening various combinations of the heads of risk, compliance, sales, and human resources, as well as the business unit general managers, to discuss capability-focused topics. Because the right people were hearing the context in real time, execution accelerated. Leaders no longer had to “sell” decisions afterward or decipher what had happened behind closed doors because they were in the room when it happened. Elevate executive team time. When too many decisions happen in private 1:1s, full executive team meetings often devolve into catch-up sessions, replaying decisions, clarifying miscommunications, or managing interpersonal fallout. That robs the team of the chance to focus on the truly enterprise-wide work only they can do. By using capability meetings to handle cross-functional execution and trade-offs at the right level, executive team time can be reserved for strategic sensemaking, cultural stewardship, and long-horizon bets. Center agendas around questions like “What are the few issues that require the full weight of this team?” or “Where does our system need realignment?” One Fortune 100 company overhauled its monthly executive team meeting after realizing that too much time was spent relitigating decisions made elsewhere. By reassigning operational decision-making to triads tied to core capabilities, they cleared the decks for their senior team to focus on systemic enterprise priorities—like cultural resilience, the innovation pipeline, scenario planning for disruptions, and global portfolio shifts. The result was faster execution at the edges and deeper strategic coherence at the center.
考点1: "vented"应该译为“透露”。 考点2: "VP"应译为“副总监”。 考点3:" functionally siloed"应译为“职能孤岛化”。 考点4: "Fragmented governance"应译为“碎片化治理 ”。 考点5: 'game the system"应译为“钻制度的空子”。 考点6: "Succession overhang"应译为“继任问题带来的潜在影响”。 考点7:"cleared the decks"应译为“扫清障碍”。
5fe02
垂类场景
广告营销
65
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Comparison between pediatric and adult acute natural cannabinoids toxicity: A 5-year retrospective study with special consideration of acute synthetic cannabinoids toxicity Abstract Acute cannabinoids toxicity is an alarming toxicological problem. The current study aimed to compare children and adults with acute natural cannabinoids toxicity and highlight cases with acute synthetic cannabinoids (SCs) toxicity. This retrospective cross-sectional study was conducted on patients with acute cannabinoids toxicity admitted to Tanta University Poison Control Center from January 2019 to December 2023. Socio-demographic, toxicological, clinical, and laboratory data were retrieved from patients’ medical records. Patients were divided into a pediatric group (≤ 18 years) and an adult group (> 18 years). Out of 106 patients, 68 were children and 38 were adults. Impaired consciousness level and bradypnea were more significantly reported in children (P < 0.001, 0.007, respectively). Low oxygen saturation, tachycardia, hypokalemia, and leukocytosis were more significantly reported in adults (P < 0.001, for each). Delay time from exposure to medical intervention and potassium level were significantly valid to predict complications in children (Adjusted odds ratio: 1.393 and 4.139, respectively). Delay time to medical intervention and oxygen saturation were significant risk factors for prolonged hospital stay in children (Adjusted odds ratio: 1.255 and 0.677, respectively). Acute SCs toxicity was observed only in four cases presented mainly with seizures, tachycardia, hypertension, tachypnea, and hypoxemia. It could be concluded that natural cannabinoids toxicity is more prevalent than SCs. Presentation of acute natural cannabinoids toxicity exhibits variations between children and adults. Delay time to medical intervention, as well as potassium and oxygen saturation levels are significant risk factors for complications and prolonged hospitalization in children. Keywords: Acute cannabis poisoning, Synthetic cannabinoids, Prediction, Outcomes 1. Introduction Over the last decades, substance abuse has emerged as a significant health issue that strongly affects all countries. As defined by the World Health Organization (WHO), substance abuse is the hazardous use of psychoactive substances, including alcohol and illicit drugs with substantial risk of morbidity and premature death [1]. As other countries, Egypt is witnessing increasing rates of substance abuse while the exact prevalence is unavailable, most probably due to substantial social stigma and underreporting [2]. Some reports highlight the magnitude of the problem; in 2016, about 6.8 % of Egyptians aged above 15 years were involved in substance abuse [3]. Furthermore, in 2022, the Egyptian Ministry of Social Solidarity estimated that about 5.9 % of the total population consumed illicit drugs [4]. Globally, cannabis ranks third among abused substances after alcohol and tobacco as it is widely cultivated and trafficked with an ancient history of both medicinal and recreational uses [5], [6] . The annual prevalence rate of cannabis consumption has been estimated to be nearly 2.5 % of the global population [7]. In Egypt, different reports are available considering the prevalence of cannabis abuse. For example, it was reported that more than 9 % of Egyptian students are using bango and 3 % are depending on hashish [8]. Despite the fact that recreational cannabis use is known to be associated with a low risk of acute toxicity, recent reports indicate an increase in the potency of cannabis preparations, potentially leading to more visits to poison control centers due to accidental acute toxicity [9]. Furthermore, accidental ingestion of cannabis by children is also reported even in developed countries especially with legalization of medical and recreational marijuana [10]. Although it is an alarming toxicological problem, data regarding acute cannabis toxicity is still limited [11]. It is worth mentioning that the term cannabinoid refers to chemical substances that interact with cannabinoid receptors, producing effects similar to those of the Cannabis plant [12]. Depending on the source of production, cannabinoids are divided into endocannabinoids, phytocannabinoids and synthetic cannabinoids. The first group (endocannabinoids) is produced by mammalians while the second group (phytocannabinoids) is extracted from the cannabis plant and includes the most psychoactive Δ9-tetrahydrocannabinol (Δ9-THC) and non-psychoactive component cannabidiol. The third group involves a number of synthetic cannabinoids (SCs) which are sold for recreational use [13]. Toxicity from natural cannabinoids (marijuana) including hashish and bango can result from inhalation, smoking, or ingestion [14], [15], [16]. Manifestations are diverse depending on the patient’s age, cannabis potency, route of exposure, and combination with other psychoactive substances [17]. Basic treatment of acute cannabis toxicity is mainly supportive to keep a patent airway with adequate ventilation and circulation. Children presented with apnea and risk of aspiration need to be intubated and mechanically ventilated. However, those who are presenting with lethargy, rapid electrolytes and arterial blood gases, as well as random blood sugar measurement is considered. For patients with mild toxicity, benzodiazepine may be enough to control anxiety and agitation [18]. Regarding SCs, they are designer drugs with variable structure and potencies that mimic the psychoactive properties of Δ9-THC. In Egypt, they are traded under different names including strox and voodoo [16]. In 2014, they were added to Schedule No. 1 of the Egyptian Drugs Act [19]. Synthetic cannabinoids acquired their popularity because they are easily manufactured, so they are more accessible at low prices compared to cannabis. Additionally, they are not detected by standard toxicology screening for natural cannabis [15]. Despite the cannabomimetic effects of SCs, their side effects cannot be expected and may be more serious than the natural cannabis [20]. Acute toxicity could be due to Δ9-THC and additives like anticholinergic agents and ketamine [21]. However, acute strox intoxication is manifested by visual and auditory hallucinations, fright, speech impairments, and paranoia leading to aggressive behavior [22], [15]. It may also cause loss of concentration, delirium, vomiting, and fainting, as well as extreme fear of death and fatal convulsions [23]. Pupillary dilatation, tachycardia, flushed dry skin, and other anticholinergic toxidrome were also reported. In severe cases, lethal cardiovascular collapse and deep coma could occur [24]. Supportive and symptomatic care is the main line considered in the treatment of acute strox toxicity [25]. Considering that age could affect the characteristics of acute toxicity and the introduction of SCs into the market of substance abuse, this study aimed to compare clinical and laboratory findings of acute natural cannabinoids toxicity in the adults and pediatric group, together with highlighting cases presented with acute SCs toxicity. 2. Patients and methods 2.1. Study design and settings The current retrospective cross-sectional study was conducted on all patients with acute cannabis or SCs toxicity who were admitted to Tanta University Poison Control Center (TUPCC) throughout five years from January 2019 to December 2023. 2.2. Ethical considerations The current study was approved by our institution's ethics committee (approval code: 36264PR708/5/24) and followed the declaration of Helsinki. To secure confidentiality, patients’ data were coded for anonymous statistical analysis. The requirement for written informed consent was waived due to the retrospective nature of the study. 2.3. Eligibility criteria All patients, males and females, with acute cannabis or SCs toxicity were included in the study. Diagnosis was based on definite history as reported from the patients or witnesses besides clinical assessment of the attending physician and urine screening test [26]. On the other hand, patients with known medical illnesses such as cardiovascular or neuropsychological diseases, renal or hepatic impairment, respiratory disorders such as asthma or chronic obstructive pulmonary disorders, and diabetes were excluded. Additionally, the current study excluded patients with mixed intoxication or those who received medical intervention before admission. Pregnant and lactating females were also excluded. 2.4. Data collection The following variables were extracted from the medical records of each patient: 1.Socio-demographics: age, sex, and residence. 2.History of pre-existing medical disorders. 3.Toxicological data: name of substance, route, manner, and place of exposure, and delay time from exposure to medical intervention. 4.Findings of clinical examination: •Vital signs: systolic blood pressure (SBP), diastolic blood pressure (DBP), pulse rate (PR), respiratory rate (RR), temperature, and oxygen (O2) saturation. •Level of consciousness using the Glasgow Coma Scale (GCS). •Systemic examination: head & neck, neurological, chest, abdomen, skin, and extremities examination •Electrocardiographic (ECG) findings 1.Results of laboratory investigations: •Arterial blood gases (ABG): pH, partial arterial carbon dioxide pressure (PaCO2), partial arterial oxygen pressure (PaO2). •Serum electrolytes: serum sodium (Na) and potassium (K) levels •Serum random blood glucose level •Glucose potassium (Glu/K) ratio was obtained by dividing serum random blood glucose level over serum potassium level •Renal functions: blood urea nitrogen and serum creatinine. •Liver enzymes: aspartate aminotransferase (AST) and alanine aminotransferase (ALT). 2.Outcome measures: mortality, intensive care unit (ICU) admission, and duration of hospital stay. Patients were treated according to guidelines including emergency and supportive measures if indicated. Patients were divided according to their age into two groups: the pediatric group included patients whose age was 18 years or less and the adult group where the age of patients was above 18 years. 2.5. Statistical analysis All data were tabulated and analyzed by the statistical package for the social sciences software program, IBM SPSS Statistics for Windows, version 27 (IBM Corp., Armonk, N.Y., USA). Categorical data were presented as numbers and percentages, while numerical data were first tested for normality by the Shapiro-Wilk test. Normally distributed numerical data were expressed as the mean ± standard deviation, while skewed numerical data were represented as median and interquartile range (IQR: 25th −75th percentiles). Comparisons between the studied groups were done by applying Pearson’s Chi-Square test for categorical data. When more than 20 % of cells have expected count of less than 5 Fisher’s Exact and Fisher-Freeman-Halton Exact tests was adopted instead of Pearson’s Chi-Square test. Normally distributed numerical data were compared by the Independent Samples T-test. Alternatively, Mann-Whitney U test was performed to compare skewed data. For determining risk factors of complicated outcome and prolonged hospital stay among the studied children group, univariable and multivariable logistic regression analyses were performed. As regards multivariable logistic regression, all variables that showed significant results in the bivariate analysis were selected as covariates. Results were presented as adjusted odds ratio (AOR) and 95 % confidence interval. P < 0.05 was considered statistically significant.
Comparison between pediatric and adult acute natural cannabinoids toxicity: A 5-year retrospective study with special consideration of acute synthetic cannabinoids toxicity Abstract Acute cannabinoids toxicity is an alarming toxicological problem. The current study aimed to compare children and adults with acute natural cannabinoids toxicity and highlight cases with acute synthetic cannabinoids (SCs) toxicity. This retrospective cross-sectional study was conducted on patients with acute cannabinoids toxicity admitted to Tanta University Poison Control Center from January 2019 to December 2023. Socio-demographic, toxicological, clinical, and laboratory data were retrieved from patients’ medical records. Patients were divided into a pediatric group (≤ 18 years) and an adult group (> 18 years). Out of 106 patients, 68 were children and 38 were adults. Impaired consciousness level and bradypnea were more significantly reported in children (P < 0.001, 0.007, respectively). Low oxygen saturation, tachycardia, hypokalemia, and leukocytosis were more significantly reported in adults (P < 0.001, for each). Delay time from exposure to medical intervention and potassium level were significantly valid to predict complications in children (Adjusted odds ratio: 1.393 and 4.139, respectively). Delay time to medical intervention and oxygen saturation were significant risk factors for prolonged hospital stay in children (Adjusted odds ratio: 1.255 and 0.677, respectively). Acute SCs toxicity was observed only in four cases presented mainly with seizures, tachycardia, hypertension, tachypnea, and hypoxemia. It could be concluded that natural cannabinoids toxicity is more prevalent than SCs. Presentation of acute natural cannabinoids toxicity exhibits variations between children and adults. Delay time to medical intervention, as well as potassium and oxygen saturation levels are significant risk factors for complications and prolonged hospitalization in children. Keywords: Acute cannabis poisoning, Synthetic cannabinoids, Prediction, Outcomes 1. Introduction Over the last decades, substance abuse has emerged as a significant health issue that strongly affects all countries. As defined by the World Health Organization (WHO), substance abuse is the hazardous use of psychoactive substances, including alcohol and illicit drugs with substantial risk of morbidity and premature death [1]. As other countries, Egypt is witnessing increasing rates of substance abuse while the exact prevalence is unavailable, most probably due to substantial social stigma and underreporting [2]. Some reports highlight the magnitude of the problem; in 2016, about 6.8 % of Egyptians aged above 15 years were involved in substance abuse [3]. Furthermore, in 2022, the Egyptian Ministry of Social Solidarity estimated that about 5.9 % of the total population consumed illicit drugs [4]. Globally, cannabis ranks third among abused substances after alcohol and tobacco as it is widely cultivated and trafficked with an ancient history of both medicinal and recreational uses [5], [6] . The annual prevalence rate of cannabis consumption has been estimated to be nearly 2.5 % of the global population [7]. In Egypt, different reports are available considering the prevalence of cannabis abuse. For example, it was reported that more than 9 % of Egyptian students are using bango and 3 % are depending on hashish [8]. Despite the fact that recreational cannabis use is known to be associated with a low risk of acute toxicity, recent reports indicate an increase in the potency of cannabis preparations, potentially leading to more visits to poison control centers due to accidental acute toxicity [9]. Furthermore, accidental ingestion of cannabis by children is also reported even in developed countries especially with legalization of medical and recreational marijuana [10]. Although it is an alarming toxicological problem, data regarding acute cannabis toxicity is still limited [11]. It is worth mentioning that the term cannabinoid refers to chemical substances that interact with cannabinoid receptors, producing effects similar to those of the Cannabis plant [12]. Depending on the source of production, cannabinoids are divided into endocannabinoids, phytocannabinoids and synthetic cannabinoids. The first group (endocannabinoids) is produced by mammalians while the second group (phytocannabinoids) is extracted from the cannabis plant and includes the most psychoactive Δ9-tetrahydrocannabinol (Δ9-THC) and non-psychoactive component cannabidiol. The third group involves a number of synthetic cannabinoids (SCs) which are sold for recreational use [13]. Toxicity from natural cannabinoids (marijuana) including hashish and bango can result from inhalation, smoking, or ingestion [14], [15], [16]. Manifestations are diverse depending on the patient’s age, cannabis potency, route of exposure, and combination with other psychoactive substances [17]. Basic treatment of acute cannabis toxicity is mainly supportive to keep a patent airway with adequate ventilation and circulation. Children presented with apnea and risk of aspiration need to be intubated and mechanically ventilated. However, those who are presenting with lethargy, rapid electrolytes and arterial blood gases, as well as random blood sugar measurement is considered. For patients with mild toxicity, benzodiazepine may be enough to control anxiety and agitation [18]. Regarding SCs, they are designer drugs with variable structure and potencies that mimic the psychoactive properties of Δ9-THC. In Egypt, they are traded under different names including strox and voodoo [16]. In 2014, they were added to Schedule No. 1 of the Egyptian Drugs Act [19]. Synthetic cannabinoids acquired their popularity because they are easily manufactured, so they are more accessible at low prices compared to cannabis. Additionally, they are not detected by standard toxicology screening for natural cannabis [15]. Despite the cannabomimetic effects of SCs, their side effects cannot be expected and may be more serious than the natural cannabis [20]. Acute toxicity could be due to Δ9-THC and additives like anticholinergic agents and ketamine [21]. However, acute strox intoxication is manifested by visual and auditory hallucinations, fright, speech impairments, and paranoia leading to aggressive behavior [22], [15]. It may also cause loss of concentration, delirium, vomiting, and fainting, as well as extreme fear of death and fatal convulsions [23]. Pupillary dilatation, tachycardia, flushed dry skin, and other anticholinergic toxidrome were also reported. In severe cases, lethal cardiovascular collapse and deep coma could occur [24]. Supportive and symptomatic care is the main line considered in the treatment of acute strox toxicity [25]. Considering that age could affect the characteristics of acute toxicity and the introduction of SCs into the market of substance abuse, this study aimed to compare clinical and laboratory findings of acute natural cannabinoids toxicity in the adults and pediatric group, together with highlighting cases presented with acute SCs toxicity. 2. Patients and methods 2.1. Study design and settings The current retrospective cross-sectional study was conducted on all patients with acute cannabis or SCs toxicity who were admitted to Tanta University Poison Control Center (TUPCC) throughout five years from January 2019 to December 2023. 2.2. Ethical considerations The current study was approved by our institution's ethics committee (approval code: 36264PR708/5/24) and followed the declaration of Helsinki. To secure confidentiality, patients’ data were coded for anonymous statistical analysis. The requirement for written informed consent was waived due to the retrospective nature of the study. 2.3. Eligibility criteria All patients, males and females, with acute cannabis or SCs toxicity were included in the study. Diagnosis was based on definite history as reported from the patients or witnesses besides clinical assessment of the attending physician and urine screening test [26]. On the other hand, patients with known medical illnesses such as cardiovascular or neuropsychological diseases, renal or hepatic impairment, respiratory disorders such as asthma or chronic obstructive pulmonary disorders, and diabetes were excluded. Additionally, the current study excluded patients with mixed intoxication or those who received medical intervention before admission. Pregnant and lactating females were also excluded. 2.4. Data collection The following variables were extracted from the medical records of each patient: 1.Socio-demographics: age, sex, and residence. 2.History of pre-existing medical disorders. 3.Toxicological data: name of substance, route, manner, and place of exposure, and delay time from exposure to medical intervention. 4.Findings of clinical examination: •Vital signs: systolic blood pressure (SBP), diastolic blood pressure (DBP), pulse rate (PR), respiratory rate (RR), temperature, and oxygen (O2) saturation. •Level of consciousness using the Glasgow Coma Scale (GCS). •Systemic examination: head & neck, neurological, chest, abdomen, skin, and extremities examination •Electrocardiographic (ECG) findings 1.Results of laboratory investigations: •Arterial blood gases (ABG): pH, partial arterial carbon dioxide pressure (PaCO2), partial arterial oxygen pressure (PaO2). •Serum electrolytes: serum sodium (Na) and potassium (K) levels •Serum random blood glucose level •Glucose potassium (Glu/K) ratio was obtained by dividing serum random blood glucose level over serum potassium level •Renal functions: blood urea nitrogen and serum creatinine. •Liver enzymes: aspartate aminotransferase (AST) and alanine aminotransferase (ALT). 2.Outcome measures: mortality, intensive care unit (ICU) admission, and duration of hospital stay. Patients were treated according to guidelines including emergency and supportive measures if indicated. Patients were divided according to their age into two groups: the pediatric group included patients whose age was 18 years or less and the adult group where the age of patients was above 18 years. 2.5. Statistical analysis All data were tabulated and analyzed by the statistical package for the social sciences software program, IBM SPSS Statistics for Windows, version 27 (IBM Corp., Armonk, N.Y., USA). Categorical data were presented as numbers and percentages, while numerical data were first tested for normality by the Shapiro-Wilk test. Normally distributed numerical data were expressed as the mean ± standard deviation, while skewed numerical data were represented as median and interquartile range (IQR: 25th −75th percentiles). Comparisons between the studied groups were done by applying Pearson’s Chi-Square test for categorical data. When more than 20 % of cells have expected count of less than 5 Fisher’s Exact and Fisher-Freeman-Halton Exact tests was adopted instead of Pearson’s Chi-Square test. Normally distributed numerical data were compared by the Independent Samples T-test. Alternatively, Mann-Whitney U test was performed to compare skewed data. For determining risk factors of complicated outcome and prolonged hospital stay among the studied children group, univariable and multivariable logistic regression analyses were performed. As regards multivariable logistic regression, all variables that showed significant results in the bivariate analysis were selected as covariates. Results were presented as adjusted odds ratio (AOR) and 95 % confidence interval. P < 0.05 was considered statistically significant.
考点 1:"Acute cannabinoids toxicity" 应译为 "急性大麻素毒性" 考点 2:"Impaired consciousness level" 应译为 "意识障碍" 考点 3:"leukocytosis" 应译为 "白细胞增多症" 考点 4:"Adjusted odds ratio" 应译为 "校正后比值比" 考点 5:"Outcomes" 应译为 "临床结局" 考点 6:"morbidity and premature death" 应译为 "发病率和过早死亡" 考点 7:"prevalence" 应译为 "流行率" 考点 8:"bango and hashish" 应译为 "班戈和哈希什" 考点 9:"poison control centers" 应译为 "中毒控制中心" 考点 10:"patent airway" 应译为 "保持气道通畅" 考点 11:"intubated and mechanically ventilated" 应译为 "气管插管和机械通气" 考点 12:"designer drugs" 应译为 "设计毒品" 考点 13:"strox and voodoo" 应译为 "斯特罗 和 巫毒" 考点 14:"cannabomimetic effects" 应译为 "拟大麻素作用" 考点 15:"anticholinergic toxidrome" 应译为 "抗胆碱能毒性症候群" 考点 33:"lethal cardiovascular collapse" 应译为 "致死性心血管衰竭" 考点 34:"waived" 应译为 "豁免" 考点 35:"Eligibility criteria" 应译为 "纳入标准" 考点 36:"mixed intoxication" 应译为 "混合中毒"
6108f
学术论文
自然科学
90
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 核心观点:“房地产”是宏观经济研究绕不开的“锚”,在经济增长、产业发展、金融市场、居民资产负债表、财政税收等领域都扮演过重要的角色。2021年以来,宏观经济进入“去地产化”阶段,房地产市场景气变化对增长、就业、税收、物价、系统性风险等领域带来了不同程度的“扰动”,市场存在“地产见顶”的认识,也有所谓的“唯地产论”论调甚嚣尘上。24年924后高层提出“止跌回稳”的政策基调,我们也判断房地产市场“最悲观的阶段”已经过去、进入“见龙在田”阶段,房地产市场预期有所改观,25年房地产在经历了年初的小阳春后有所降温,二季度以来与地产相关的悲观叙事再度浮现,包括但不限于担心头部房企债务压力、金融体系的系统性风险,甚至是经济增长前景等等。 在此背景下,我们认为不能以“悲观叙事”就否定房地产领域的积极变化和投资机会,不破不立,风险中暗藏着变化,变化中孕育着生机,我们希望通过“地产进化论”系列报告对房地产领域的系列问题进行探讨,为投资者掘金“新机”。 本篇报告作为开篇,我们希望以长周期视角出发,从供给和需求两个方面梳理未来存量、增量的变化。可以看到,我国房地产供给端的“绝对增量”高峰已经过去,从“跑马圈地”的增量时代进入存量时代,居民城镇住房也从“有没有”逐步转向“好不好”和“够不够宽敞”的改善阶段。需求端,在总量“承压”的背景下,也仍有结构性的机会,一是市民化和城镇化率提升带来增量需求;二是我国老龄化程度逐步提升,意味着养老住房的需求也在逐步提升;三是房地产领域也呈现出“品质消费时代”特征,在主力购房人群规模边际下降的背景下,高品质的“好房子”需求显著提升,这也决定了房地产市场会出现“存量供给淘汰出清”并催生各类“更新”需求。 一、地产供给 增量视角:在供给端,我们认为最直接、最具代表性的切入点就是“户均住宅套数”或“套户比”这一指标:我们测算发现,在城镇家庭户口径下,2020年商品住宅套户比达到0.63、住宅套户比达到1.06(首次突破1)。进一步,截至2023年,商品住宅套户比达到0.70、住宅套户比达到1.17。此外,考虑到“户均住宅套数”这一指标仍无官方口径披露,我们的测算值仅为一个简单的参考,人口普查中的“人均住房间数”和“人均住房建筑面积”则是另两个值得重点关注的长周期指标。2020年第七次人口普查时人均住房间数达到1.0间/人,这意味着平均每位城镇居民已基本拥有一个独立房间,居住分隔水平相较20年前提升了约30%。人均住房建筑面积也在同步扩大,到2020年已经增长至38.62平方米/人,二十年累计增长超过70%。 存量视角:从增量层面出发,我们认为“人均住宅新开工套数”是供给层面最直观的增量信号,也是对存量供给视角的重要补充,它决定了未来1-3年可供入市的新房供给节奏。从历史数据看,我国每千人的人均住宅新开工套数经历了典型的“快速放量-阶段性见顶-有序回落”三个阶段。到2023年,全国千人新开工套数已降至约4.9套,2024年进一步回落至3.8套;城镇人口口径同样从高点大幅回落到2023年的7.5套左右,2024年降至5.7套/千人左右。换言之,我国房地产供给端的“绝对增量”高峰已经过去,开发节奏正与人口增长和城镇化红利的边际收窄相适应,新增供给趋于理性。 二、地产需求 地产需求拆分:从长期视角看,房地产的需求亦可以分为存量与增量两大层面,我们认为在结构上可进一步拆解为四大主要来源:1)一是城镇化带来的新增购房需求,主要由农业转移人口进城购房形成;2)二是城镇主力购房人群的刚性与改善性需求,核心集中在25-44岁人群,其中25-34岁年龄段以首套购房为主,35-44岁人群则以改善性换房为主;3)三是养老住房需求,主要对应老龄人口规模增加所带来的以房换房或适老化改善需求;4)四是投机性购房需求,通常在房价预期持续上涨阶段出现,对市场热度有显著放大效应。 存量视角: 1)从存量角度看,城镇化是过去二十年推动中国住房需求增长的最大变量之一,累计带来数亿农民进城落户、购房、安居的需求。而从国际比较来看,中国的城镇化水平虽然已明显接近中高收入经济体的平均线,但与发达经济体仍有一定差距; 2)在城镇化进程逐步迈入高位平台期的同时,另一项与住房需求息息相关的人口结构变量,也正悄然发生变化,即适龄购房人口群体的演变; 3)我国65岁及以上人口占比已超过15%,进入中度老龄化社会。从趋势上看,预计老龄人口占比仍将持续上升,未来一段时期内,养老相关的居住需求将逐步上升为住房市场中的一个重要分支; 4)投机性需求的存量部分,往往体现在房价预期带来的多套房持有行为和“囤房”现象。在过去房价持续上涨的周期中,部分家庭或机构投资者积累了多套住房,形成了对存量市场的额外“供给约束”。 增量视角: 对四类住房需求的展望: 1)从整体上看,当前来看我国常住人口城镇化率已达到近65%,处于较高水平,但仍有进一步上行的空间,中西部地区和部分城市群外围的农村安居落户仍在继续,这背后会带来两种增量购房需求:一是市民化;二是常住人口城镇化率继续提升所带来的阶段性购房需求;从结构上看,“城镇化”对不同区域的影响存在差异; 2)25-44岁人群历来是住房市场最核心的消费群体。25-34岁以首套刚需为主,35-44岁则以改善性换房为主。尽管适龄人口总体呈下降趋势,但人口净流入城市、经济活跃地区仍具备年轻人口积聚能力。此外,代际财富传承、房贷政策优化、消费观念转变等因素可能延缓购房节奏但不削弱购房意愿; 3)随着人口老龄化持续推进,养老住房需求正在从边缘走向主流。老年人更多以“以房换房”或局部改善方式参与市场,核心诉求是改善适老条件、优化生活环境或靠近子女生活圈。未来,“银发经济”发展和养老住宅细分业态的成熟,将形成结构性增量; 4)投机性购房需求具有典型顺周期特征,过去曾在房价上涨周期推动市场升温,但当前政策导向下其影响力大幅下降。未来在“弱周期+低杠杆”背景下,投机性需求的空间将更受约束,但在市场预期转向时仍可能阶段性抬头; 5)在房地产市场由增量扩张转向存量优化的阶段,“好房子”正逐渐成为各类购房需求的重要共同指向。无论是主力购房人群的改善性需求、老年人口的养老置换,还是投机性需求,均表现出对高品质住房的高度偏好。“好房子”不仅是改善居住体验的目标,也正成为房地产需求侧的关键锚点。
核心观点:“房地产”是宏观经济研究绕不开的“锚”,在经济增长、产业发展、金融市场、居民资产负债表、财政税收等领域都扮演过重要的角色。2021年以来,宏观经济进入“去地产化”阶段,房地产市场景气变化对增长、就业、税收、物价、系统性风险等领域带来了不同程度的“扰动”,市场存在“地产见顶”的认识,也有所谓的“唯地产论”论调甚嚣尘上。24年924后高层提出“止跌回稳”的政策基调,我们也判断房地产市场“最悲观的阶段”已经过去、进入“见龙在田”阶段,房地产市场预期有所改观,25年房地产在经历了年初的小阳春后有所降温,二季度以来与地产相关的悲观叙事再度浮现,包括但不限于担心头部房企债务压力、金融体系的系统性风险,甚至是经济增长前景等等。 在此背景下,我们认为不能以“悲观叙事”就否定房地产领域的积极变化和投资机会,不破不立,风险中暗藏着变化,变化中孕育着生机,我们希望通过“地产进化论”系列报告对房地产领域的系列问题进行探讨,为投资者掘金“新机”。 本篇报告作为开篇,我们希望以长周期视角出发,从供给和需求两个方面梳理未来存量、增量的变化。可以看到,我国房地产供给端的“绝对增量”高峰已经过去,从“跑马圈地”的增量时代进入存量时代,居民城镇住房也从“有没有”逐步转向“好不好”和“够不够宽敞”的改善阶段。需求端,在总量“承压”的背景下,也仍有结构性的机会,一是市民化和城镇化率提升带来增量需求;二是我国老龄化程度逐步提升,意味着养老住房的需求也在逐步提升;三是房地产领域也呈现出“品质消费时代”特征,在主力购房人群规模边际下降的背景下,高品质的“好房子”需求显著提升,这也决定了房地产市场会出现“存量供给淘汰出清”并催生各类“更新”需求。 一、地产供给 增量视角:在供给端,我们认为最直接、最具代表性的切入点就是“户均住宅套数”或“套户比”这一指标:我们测算发现,在城镇家庭户口径下,2020年商品住宅套户比达到0.63、住宅套户比达到1.06(首次突破1)。进一步,截至2023年,商品住宅套户比达到0.70、住宅套户比达到1.17。此外,考虑到“户均住宅套数”这一指标仍无官方口径披露,我们的测算值仅为一个简单的参考,人口普查中的“人均住房间数”和“人均住房建筑面积”则是另两个值得重点关注的长周期指标。2020年第七次人口普查时人均住房间数达到1.0间/人,这意味着平均每位城镇居民已基本拥有一个独立房间,居住分隔水平相较20年前提升了约30%。人均住房建筑面积也在同步扩大,到2020年已经增长至38.62平方米/人,二十年累计增长超过70%。 存量视角:从增量层面出发,我们认为“人均住宅新开工套数”是供给层面最直观的增量信号,也是对存量供给视角的重要补充,它决定了未来1-3年可供入市的新房供给节奏。从历史数据看,我国每千人的人均住宅新开工套数经历了典型的“快速放量-阶段性见顶-有序回落”三个阶段。到2023年,全国千人新开工套数已降至约4.9套,2024年进一步回落至3.8套;城镇人口口径同样从高点大幅回落到2023年的7.5套左右,2024年降至5.7套/千人左右。换言之,我国房地产供给端的“绝对增量”高峰已经过去,开发节奏正与人口增长和城镇化红利的边际收窄相适应,新增供给趋于理性。 二、地产需求 地产需求拆分:从长期视角看,房地产的需求亦可以分为存量与增量两大层面,我们认为在结构上可进一步拆解为四大主要来源:1)一是城镇化带来的新增购房需求,主要由农业转移人口进城购房形成;2)二是城镇主力购房人群的刚性与改善性需求,核心集中在25-44岁人群,其中25-34岁年龄段以首套购房为主,35-44岁人群则以改善性换房为主;3)三是养老住房需求,主要对应老龄人口规模增加所带来的以房换房或适老化改善需求;4)四是投机性购房需求,通常在房价预期持续上涨阶段出现,对市场热度有显著放大效应。 存量视角: 1)从存量角度看,城镇化是过去二十年推动中国住房需求增长的最大变量之一,累计带来数亿农民进城落户、购房、安居的需求。而从国际比较来看,中国的城镇化水平虽然已明显接近中高收入经济体的平均线,但与发达经济体仍有一定差距; 2)在城镇化进程逐步迈入高位平台期的同时,另一项与住房需求息息相关的人口结构变量,也正悄然发生变化,即适龄购房人口群体的演变; 3)我国65岁及以上人口占比已超过15%,进入中度老龄化社会。从趋势上看,预计老龄人口占比仍将持续上升,未来一段时期内,养老相关的居住需求将逐步上升为住房市场中的一个重要分支; 4)投机性需求的存量部分,往往体现在房价预期带来的多套房持有行为和“囤房”现象。在过去房价持续上涨的周期中,部分家庭或机构投资者积累了多套住房,形成了对存量市场的额外“供给约束”。 增量视角: 对四类住房需求的展望: 1)从整体上看,当前来看我国常住人口城镇化率已达到近65%,处于较高水平,但仍有进一步上行的空间,中西部地区和部分城市群外围的农村安居落户仍在继续,这背后会带来两种增量购房需求:一是市民化;二是常住人口城镇化率继续提升所带来的阶段性购房需求;从结构上看,“城镇化”对不同区域的影响存在差异; 2)25-44岁人群历来是住房市场最核心的消费群体。25-34岁以首套刚需为主,35-44岁则以改善性换房为主。尽管适龄人口总体呈下降趋势,但人口净流入城市、经济活跃地区仍具备年轻人口积聚能力。此外,代际财富传承、房贷政策优化、消费观念转变等因素可能延缓购房节奏但不削弱购房意愿; 3)随着人口老龄化持续推进,养老住房需求正在从边缘走向主流。老年人更多以“以房换房”或局部改善方式参与市场,核心诉求是改善适老条件、优化生活环境或靠近子女生活圈。未来,“银发经济”发展和养老住宅细分业态的成熟,将形成结构性增量; 4)投机性购房需求具有典型顺周期特征,过去曾在房价上涨周期推动市场升温,但当前政策导向下其影响力大幅下降。未来在“弱周期+低杠杆”背景下,投机性需求的空间将更受约束,但在市场预期转向时仍可能阶段性抬头; 5)在房地产市场由增量扩张转向存量优化的阶段,“好房子”正逐渐成为各类购房需求的重要共同指向。无论是主力购房人群的改善性需求、老年人口的养老置换,还是投机性需求,均表现出对高品质住房的高度偏好。“好房子”不仅是改善居住体验的目标,也正成为房地产需求侧的关键锚点。
考点1:见龙在田 推荐译为 a new begining,因为需要将这个词背后的含义翻译出来,表示新的开始 考点2:小阳春 推荐译为 mini-boom,因为需要表达经济回暖的状况 考点3:跑马圈地 推荐译为 land grabbing,因为需要表现出原文中对土地抢夺无序的状况 考点4:适老化改善需求 推荐译为 age-friendly upgrade demand
61467
学术论文
社会科学
159
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 宝钗自到荣府,因她美丽端方,行为豁达,入境随俗,不像黛玉那样清高自负,所以深得下人的喜爱,就是那些小丫头们也乐于和她说笑。黛玉心中有些不忿,她还不知。宝玉天性爱和女孩儿厮混,也看不出来,对她们都如姐妹,没有亲疏远近之分。 东边宁府花园里梅花盛开,贾珍的妻子尤氏置办酒席,带着贾蓉夫妻去请贾母、邢夫人、王夫人等过来赏花。贾母等早饭后都过来,到会芳园游玩,先茶后酒。晌午时,宝玉困乏,想睡午觉。贾蓉的妻子秦氏就说:“我们有给宝叔收拾的房子,请老祖宗放心。”就带上宝玉一班人来到上房内间。宝玉见墙上挂着一幅鼓励人奋发勤学的《燃黎图》,还有一副对联: 世事洞明皆学问,人情练达即文章。 忙说:“快出去,快出去!”秦氏说:“宝叔嫌这里不好,就睡我屋里去吧。”一个嬷嬷说:“哪有小叔睡侄媳妇房里的?”秦氏笑着说:“哎哟,他有多大?别看宝叔和我弟弟同岁,两人站一起,还没有我弟弟高呢!”说着大家来到秦氏房中,只觉一股细细的甜香扑鼻。宝玉看墙上,挂着唐伯虎的《海棠春睡图》,两边是宋朝学士秦少游的对联: 嫩寒锁梦因春冷,芳气袭人是酒香。 再看房内摆设,没有一样不富丽堂皇。嬷嬷们服侍宝玉睡好,只留袭人、秋纹、晴雯、麝月四个丫头陪伴。秦氏出了房门,跟她的丫头们在廊檐下看猫儿打架。 宝玉睡去,恍恍惚惚只觉得跟着秦氏,来到一个地方,但见朱栏玉砌,绿树清溪,是人间罕见之处。正高兴,忽听山后有人唱歌: 春梦随云散,飞花逐水流。 寄言众儿女,何必觅闲愁。 宝玉听出是女孩儿的声音,放眼看去,见那边过来一位千娇百媚的丽人。再看她风姿翩翩,娉婷袅娜,与凡人不同,忙上前作揖,笑嘻嘻地问:“神仙姐姐,你从哪里来?到哪里去?能不能带上我?”仙姑说:“我住离恨天上,灌愁海中,乃是放春山遣香洞太虚幻境警幻仙姑,专司人间的风情月债,执掌尘世的女怨男痴。因近来风流冤孽,缠绵于此,所以来访察机会,布散相思。今日与你相逢,也非偶然。请跟我去一趟,听我新填《红楼梦》仙曲十二支。”宝玉欢喜踊跃,把秦氏忘到一边,跟着仙姑走到一个地方,见有一个石牌坊,上刻“太虚幻境”四字,两边是一副对联: 假作真时真亦假,无为有处有还无。 转过牌坊,是一座宫门,上书“孽海情天”四字,也有一副对联: 厚地高天,堪叹古今情不尽; 痴男怨女,可怜风月债难酬。 宝玉心中迷惑,猜不透什么是“古今情”,解不开什么是“风月债”。他随仙姑进入二层门内,见两边配殿都有对联,一时看不过来,只见几处写着“痴情司”、“结怨司”、“朝啼司”、“暮哭司”、“春感司”、“秋感司”。他问:“仙姑能不能领我到各司游玩一下?”仙姑说:“各司存放的是普天下所有的女子过去未来的簿册,你肉眼凡胎,不便看。”宝玉不依,一定要看。仙姑见这是“薄命司”,就答应了。门上也有对联: 春恨秋悲皆自惹,花容月貌为谁妍? 宝玉感叹着进了门,见里面摆着十多个大橱,橱门上都贴着封条,封条上写着各种字。他见一个橱上的封条上书“金陵十二钗正册”,就问是什么意思。仙姑说:“就是贵省十二个冠首女子之册。”宝玉问:“金陵极大,怎么只十二个女子?”仙姑说:“贵省女子虽然多,不过择其名气大的录下来,两边二柜又差一等,那些平常人是不必记录的。”宝玉看下面一柜,写着“金陵十二钗副册”,又一柜上写着“金陵十二钗又副册”。他打开“又副册”橱门,取出一册来,翻开一页,见上面画的不是山水人物,而是一幅水墨渲染,如同乌云浊雾,后有几行字迹: 霁月难逢,彩云易散。心比天高,身为下贱。风流灵巧招人怨。寿夭多因诽谤生,多情公子空牵念。 宝玉又翻一页,画着一簇鲜花、一床破席,也有几句言辞: 枉自温柔和顺,空云似桂如兰。 堪羡优伶有福,谁知公子无缘。 宝玉不解其意,扔了这册,开了“副册”橱门,拿出一册来,揭开一页,上画一枝桂花,下面有一干涸的池塘,荷叶枯败,写着: 根并荷花一茎香,平生遭际实堪伤。 自从两地生孤木,致使香魂返故乡。
宝钗自到荣府,因她美丽端方,行为豁达,入境随俗,不像黛玉那样清高自负,所以深得下人的喜爱,就是那些小丫头们也乐于和她说笑。黛玉心中有些不忿,她还不知。宝玉天性爱和女孩儿厮混,也看不出来,对她们都如姐妹,没有亲疏远近之分。 东边宁府花园里梅花盛开,贾珍的妻子尤氏置办酒席,带着贾蓉夫妻去请贾母、邢夫人、王夫人等过来赏花。贾母等早饭后都过来,到会芳园游玩,先茶后酒。晌午时,宝玉困乏,想睡午觉。贾蓉的妻子秦氏就说:“我们有给宝叔收拾的房子,请老祖宗放心。”就带上宝玉一班人来到上房内间。宝玉见墙上挂着一幅鼓励人奋发勤学的《燃黎图》,还有一副对联: 世事洞明皆学问,人情练达即文章。 忙说:“快出去,快出去!”秦氏说:“宝叔嫌这里不好,就睡我屋里去吧。”一个嬷嬷说:“哪有小叔睡侄媳妇房里的?”秦氏笑着说:“哎哟,他有多大?别看宝叔和我弟弟同岁,两人站一起,还没有我弟弟高呢!”说着大家来到秦氏房中,只觉一股细细的甜香扑鼻。宝玉看墙上,挂着唐伯虎的《海棠春睡图》,两边是宋朝学士秦少游的对联: 嫩寒锁梦因春冷,芳气袭人是酒香。 再看房内摆设,没有一样不富丽堂皇。嬷嬷们服侍宝玉睡好,只留袭人、秋纹、晴雯、麝月四个丫头陪伴。秦氏出了房门,跟她的丫头们在廊檐下看猫儿打架。 宝玉睡去,恍恍惚惚只觉得跟着秦氏,来到一个地方,但见朱栏玉砌,绿树清溪,是人间罕见之处。正高兴,忽听山后有人唱歌: 春梦随云散,飞花逐水流。 寄言众儿女,何必觅闲愁。 宝玉听出是女孩儿的声音,放眼看去,见那边过来一位千娇百媚的丽人。再看她风姿翩翩,娉婷袅娜,与凡人不同,忙上前作揖,笑嘻嘻地问:“神仙姐姐,你从哪里来?到哪里去?能不能带上我?”仙姑说:“我住离恨天上,灌愁海中,乃是放春山遣香洞太虚幻境警幻仙姑,专司人间的风情月债,执掌尘世的女怨男痴。因近来风流冤孽,缠绵于此,所以来访察机会,布散相思。今日与你相逢,也非偶然。请跟我去一趟,听我新填《红楼梦》仙曲十二支。”宝玉欢喜踊跃,把秦氏忘到一边,跟着仙姑走到一个地方,见有一个石牌坊,上刻“太虚幻境”四字,两边是一副对联: 假作真时真亦假,无为有处有还无。 转过牌坊,是一座宫门,上书“孽海情天”四字,也有一副对联: 厚地高天,堪叹古今情不尽; 痴男怨女,可怜风月债难酬。 宝玉心中迷惑,猜不透什么是“古今情”,解不开什么是“风月债”。他随仙姑进入二层门内,见两边配殿都有对联,一时看不过来,只见几处写着“痴情司”、“结怨司”、“朝啼司”、“暮哭司”、“春感司”、“秋感司”。他问:“仙姑能不能领我到各司游玩一下?”仙姑说:“各司存放的是普天下所有的女子过去未来的簿册,你肉眼凡胎,不便看。”宝玉不依,一定要看。仙姑见这是“薄命司”,就答应了。门上也有对联: 春恨秋悲皆自惹,花容月貌为谁妍? 宝玉感叹着进了门,见里面摆着十多个大橱,橱门上都贴着封条,封条上写着各种字。他见一个橱上的封条上书“金陵十二钗正册”,就问是什么意思。仙姑说:“就是贵省十二个冠首女子之册。”宝玉问:“金陵极大,怎么只十二个女子?”仙姑说:“贵省女子虽然多,不过择其名气大的录下来,两边二柜又差一等,那些平常人是不必记录的。”宝玉看下面一柜,写着“金陵十二钗副册”,又一柜上写着“金陵十二钗又副册”。他打开“又副册”橱门,取出一册来,翻开一页,见上面画的不是山水人物,而是一幅水墨渲染,如同乌云浊雾,后有几行字迹: 霁月难逢,彩云易散。心比天高,身为下贱。风流灵巧招人怨。寿夭多因诽谤生,多情公子空牵念。 宝玉又翻一页,画着一簇鲜花、一床破席,也有几句言辞: 枉自温柔和顺,空云似桂如兰。 堪羡优伶有福,谁知公子无缘。 宝玉不解其意,扔了这册,开了“副册”橱门,拿出一册来,揭开一页,上画一枝桂花,下面有一干涸的池塘,荷叶枯败,写着: 根并荷花一茎香,平生遭际实堪伤。 自从两地生孤木,致使香魂返故乡。
考点1:尤氏 应译为 Madam You 考点2:贾母 应译为 Grandmother Jia 考点3:邢夫人 应译为 Lady Xing 考点4:王夫人 应译为 Lady Wang 考点5:会芳园 应译为 Huifang Garden 考点6:秦氏 应译为 Madam Qin 考点7:《燃黎图》 应译为 Painting of Burning Reeds 考点8:嬷嬷 应译为 nursemaid 或 matron 考点9:《海棠春睡图》 应译为 Crabapple Blossoms and Spring Slumber 考点10:朱栏玉砌 应译为 vermilion balustrades and jade steps 考点11:离恨天、灌愁海 应译为 Heaven of Parting Sorrows、Sea of Poured Grief 考点12:太虚幻境 应译为 Illusory Land of Great Void 考点13:孽海情天 应译为 Sea of Sin and Sky of Passion 考点14:痴情司、结怨司、朝啼司、暮哭司、春感司、秋感司 应译为 Bureau of Infatuation, Bureau of Resentment, Bureau of Morning Lament, Bureau of Evening Weeping, Bureau of Spring Sentiment, Bureau of Autumn Sentiment 考点15:薄命司 应译为 Registry of the Doomed Maidens 或 Office of Ill-Fated Beauties 考点16:金陵十二钗正册 应译为 Primary Register of the Twelve Beauties of Jinling 考点17:金陵十二钗副册 应译为 Secondary Register of the Twelve Beauties of Jinling 考点18:金陵十二钗又副册 应译为 Supplementary Register of the Twelve Beauties of Jinling
616be
文学艺术
小说
79
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: The U.S. dollar is the world’s primary reserve currency, and it is also the most widely used currency for trade and other international transactions. However, its hegemony has come into question in recent times due to geopolitical and geostrategic shifts. As a result, de-dollarization has increasingly become a substantive topic of discussion among investors, corporates and market participants more broadly. What are the potential implications of de-dollarization, and how is it playing out in global markets and trade? What is de-dollarization? In short, de-dollarization entails a significant reduction in the use of dollars in world trade and financial transactions, decreasing national, institutional and corporate demand for the greenback. “The concept of de-dollarization relates to changes in the structural demand for the dollar that would relate to its status as a reserve currency. This encompasses areas that relate to the longer-term use of the dollar, such as transactional dominance in FX volumes or commodities trade, denomination of liabilities and share in central bank FX reserves,” said Luis Oganes, head of Global Macro Research at J.P. Morgan. Importantly, this structural shift is distinct from the cyclical demand for the greenback, which is shorter term and has in recent times been driven by U.S. exceptionalism, including the relative outperformance of the U.S. equity market. “The world has become long on the dollar in recent years, but as U.S. exceptionalism erodes, it should be reasonable to expect the overhang in USD longs to diminish as well,” Oganes said. What are the causes and implications of de-dollarization? There are two main factors that could erode the dollar’s status. The first includes adverse events that undermine the perceived safety and stability of the greenback — and the U.S.’s overall standing as the world’s leading economic, political and military power. For instance, increased polarization in the U.S. could jeopardize its governance, which underpins its role as a global safe haven. Ongoing U.S. tariff policy could also cause investors to lose confidence in American assets. The second factor involves positive developments outside the U.S. that boost the credibility of alternative currencies — economic and political reforms in China, for example. “A candidate reserve currency must be perceived as safe and stable and must provide a source of liquidity that is sufficient to meet growing global demand,” said Alexander Wise, who covers Long-Term Strategy at J.P. Morgan. Fundamentally, de-dollarization could shift the balance of power among countries, and this could, in turn, reshape the global economy and markets. The impact would be most acutely felt in the U.S., where de-dollarization would likely lead to a broad depreciation and underperformance of U.S. financial assets versus the rest of the world. “For U.S. equities, outright and relative returns would be negatively impacted by divestment or reallocation away from U.S. markets and a severe loss in confidence. There would also likely be upward pressure on real yields due to the partial divestment of U.S. fixed income by investors, or the diversification or reduction of international reserve allocations,” Wise said. Global trade The U.S.’s share in global exports and output has declined over the past three decades, while China’s has increased substantially. Nonetheless, the transactional dominance of the dollar is still evident in FX volumes, trade invoicing, cross-border liabilities denomination and foreign currency debt issuance. In 2022, the greenback dominated 88% of traded FX volumes — close to record highs — while the Chinese yuan (CNY) made up just 7%, according to data from the Bank for International Settlements (BIS). Likewise, there is little sign of USD erosion in trade invoicing. “The share of USD and EUR has held steady over the past two decades at around 40–50%. While the share of CNY is increasing in China’s cross-border transactions as it moves to conduct bilateral trade in its own currency terms, it is still low from a global standpoint,” Oganes observed. The dollar has also stoutly maintained its superiority when it comes to cross-border liabilities, where its market share stands at 48%. And in foreign currency debt issuance, its share has remained constant since the global financial crisis, at around 70%. “The daylight from the euro, whose share is at 20%, is even greater on this front,” Oganes added. FX reserves On the other hand, de-dollarization is unfolding in central bank FX reserves, where the share of USD has slid to a two-decade low in tandem with its macro footprint. “However, the dollar share in FX reserves was lower in the early 1990s, so the recent decline to just under 60% is not completely out of the norm,” said Meera Chandan, co-head of Global FX Strategy at J.P. Morgan. While much of the reallocation of FX reserves has gone to CNY and other currencies, USD and EUR still dominate levels. “The CNY footprint is still very small, even if growing, and its push for bilateral invoicing is likely to keep this trend on the upswing,” Chandan noted. The main de-dollarization trend in FX reserves, however, pertains to the growing demand for gold. Seen as an alternative to heavily indebted fiat currencies, the share of gold in FX reserves has increased, led by emerging market (EM) central banks — China, Russia and Türkiye have been the largest buyers in the last decade. Overall, while the share of gold in FX reserves in EM is still low at 9%, the figure is more than double the 4% seen a decade ago; the corresponding share for DM countries is much larger at 20%. This increased demand has in turn partly driven the current bull market in gold, with prices forecast to climb toward $4,000/oz by mid-2026. Bond markets In a sign of de-dollarization in bond markets, the share of foreign ownership in the U.S. Treasury market has been declining over the last 15 years. USD assets, principally liquid Treasuries, account for the majority of allocated FX reserves. However, demand for Treasuries has stagnated among foreign official institutions, as the growth of FX reserves has slowed and the USD’s share of reserves has dropped from its recent peak. Similarly, the backdrop for foreign private demand has weakened — as yields have risen across DM government bond markets, Treasuries have become relatively less attractive. While foreign investors remain the largest constituent within the Treasury market, their share of ownership has fallen to 30% as of early 2025 — down from a peak of above 50% during the GFC. “Although foreign demand has not kept pace with the growth of the Treasury market for more than a decade, we must consider what more aggressive action could mean. Japan is the largest foreign creditor and alone holds more than $1.1 trillion Treasuries, or nearly 4% of the market. Accordingly, any significant foreign selling would be impactful, driving yields higher,” said Jay Barry, head of Global Rates Strategy at J.P. Morgan. According to estimates by J.P. Morgan Research, each 1-percentage-point decline in foreign holdings relative to GDP (or approximately $300 billion of Treasuries) would result in yields rising by more than 33 basis points (bp). “While this is not our base case, it nonetheless underscores the impact of foreign investment on risk-free rates,” Barry added. Commodity markets De-dollarization is most visible in commodity markets, where the greenback’s influence on pricing has diminished. “Today, a large and growing proportion of energy is being priced in non-dollar-denominated contracts,” said Natasha Kaneva, head of Global Commodities Strategy at J.P. Morgan. For example, due to Western sanctions, Russian oil products exported eastward and southward are being sold in the local currencies of buyers, or in the currencies of countries Russia perceives as friendly. Among buyers, India, China and Turkey are all either using or seeking alternatives to the dollar. Saudi Arabia is also considering adding yuan-denominated futures contracts in the pricing model of Saudi Arabian oil, though progress has been slow. Notably, cross-border trade settlement in yuan is gaining ground outside of oil too. Some Indian companies have started paying for Russian coal imports in yuan, even without the involvement of Chinese intermediaries. Bangladesh also recently decided to pay Russia for its 1.4 GW nuclear power plant in yuan. “The de-dollarization trend in the commodity trade is a boon for countries like India, China, Brazil, Thailand and Indonesia, which can now not only buy oil at a discount, but also pay for it with their own local currencies,” Kaneva noted. “This reduces the need for precautionary reserves of U.S. dollars, U.S. Treasuries and oil, which might in turn free up capital to be deployed in growth-boosting domestic projects.” Deposit dollarization in emerging markets At the other end of the spectrum, deposit dollarization — where a significant portion of a country’s bank deposits are denominated in the U.S. dollar instead of the local currency — is still evident in many EM countries. “The tendency of EM residents to dollarize in times of stress appears to be correlated across markets,” said Jonny Goulden, head of EM Fixed Income Strategy at J.P. Morgan. According to J.P. Morgan Research, dollar deposits have grown mostly uninterrupted over the last decade in EM, reaching around $830 billion for a sample set of 18 EM countries (excluding China, Singapore and Hong Kong). “While there are large regional divergences in deposit dollarization across EM, all regions are more dollarized now than they were a decade ago,” Goulden noted. Latin America is the most dollarized region, with an aggregate dollarization rate of 19.1%. EMEA’s rate stands at 15.2%, while Asia (excluding China, Singapore and Hong Kong) has the lowest rate at 9.7%. China is the exception, as its dollarization rate has been persistently falling since 2017. “This is not surprising, as this was around the time when U.S.–China relations began shifting into their current state, marked by the trade war and growing diplomatic, security and geopolitical tensions,” Goulden said. “This suggests that China, alongside progress on de-dollarizing its own cross-border transactions, has effectively been de-dollarizing the deposits of Chinese residents, adding another dimension to its efforts to separate from U.S. dominance.”
The U.S. dollar is the world’s primary reserve currency, and it is also the most widely used currency for trade and other international transactions. However, its hegemony has come into question in recent times due to geopolitical and geostrategic shifts. As a result, de-dollarization has increasingly become a substantive topic of discussion among investors, corporates and market participants more broadly. What are the potential implications of de-dollarization, and how is it playing out in global markets and trade? What is de-dollarization? In short, de-dollarization entails a significant reduction in the use of dollars in world trade and financial transactions, decreasing national, institutional and corporate demand for the greenback. “The concept of de-dollarization relates to changes in the structural demand for the dollar that would relate to its status as a reserve currency. This encompasses areas that relate to the longer-term use of the dollar, such as transactional dominance in FX volumes or commodities trade, denomination of liabilities and share in central bank FX reserves,” said Luis Oganes, head of Global Macro Research at J.P. Morgan. Importantly, this structural shift is distinct from the cyclical demand for the greenback, which is shorter term and has in recent times been driven by U.S. exceptionalism, including the relative outperformance of the U.S. equity market. “The world has become long on the dollar in recent years, but as U.S. exceptionalism erodes, it should be reasonable to expect the overhang in USD longs to diminish as well,” Oganes said. What are the causes and implications of de-dollarization? There are two main factors that could erode the dollar’s status. The first includes adverse events that undermine the perceived safety and stability of the greenback — and the U.S.’s overall standing as the world’s leading economic, political and military power. For instance, increased polarization in the U.S. could jeopardize its governance, which underpins its role as a global safe haven. Ongoing U.S. tariff policy could also cause investors to lose confidence in American assets. The second factor involves positive developments outside the U.S. that boost the credibility of alternative currencies — economic and political reforms in China, for example. “A candidate reserve currency must be perceived as safe and stable and must provide a source of liquidity that is sufficient to meet growing global demand,” said Alexander Wise, who covers Long-Term Strategy at J.P. Morgan. Fundamentally, de-dollarization could shift the balance of power among countries, and this could, in turn, reshape the global economy and markets. The impact would be most acutely felt in the U.S., where de-dollarization would likely lead to a broad depreciation and underperformance of U.S. financial assets versus the rest of the world. “For U.S. equities, outright and relative returns would be negatively impacted by divestment or reallocation away from U.S. markets and a severe loss in confidence. There would also likely be upward pressure on real yields due to the partial divestment of U.S. fixed income by investors, or the diversification or reduction of international reserve allocations,” Wise said. Global trade The U.S.’s share in global exports and output has declined over the past three decades, while China’s has increased substantially. Nonetheless, the transactional dominance of the dollar is still evident in FX volumes, trade invoicing, cross-border liabilities denomination and foreign currency debt issuance. In 2022, the greenback dominated 88% of traded FX volumes — close to record highs — while the Chinese yuan (CNY) made up just 7%, according to data from the Bank for International Settlements (BIS). Likewise, there is little sign of USD erosion in trade invoicing. “The share of USD and EUR has held steady over the past two decades at around 40–50%. While the share of CNY is increasing in China’s cross-border transactions as it moves to conduct bilateral trade in its own currency terms, it is still low from a global standpoint,” Oganes observed. The dollar has also stoutly maintained its superiority when it comes to cross-border liabilities, where its market share stands at 48%. And in foreign currency debt issuance, its share has remained constant since the global financial crisis, at around 70%. “The daylight from the euro, whose share is at 20%, is even greater on this front,” Oganes added. FX reserves On the other hand, de-dollarization is unfolding in central bank FX reserves, where the share of USD has slid to a two-decade low in tandem with its macro footprint. “However, the dollar share in FX reserves was lower in the early 1990s, so the recent decline to just under 60% is not completely out of the norm,” said Meera Chandan, co-head of Global FX Strategy at J.P. Morgan. While much of the reallocation of FX reserves has gone to CNY and other currencies, USD and EUR still dominate levels. “The CNY footprint is still very small, even if growing, and its push for bilateral invoicing is likely to keep this trend on the upswing,” Chandan noted. The main de-dollarization trend in FX reserves, however, pertains to the growing demand for gold. Seen as an alternative to heavily indebted fiat currencies, the share of gold in FX reserves has increased, led by emerging market (EM) central banks — China, Russia and Türkiye have been the largest buyers in the last decade. Overall, while the share of gold in FX reserves in EM is still low at 9%, the figure is more than double the 4% seen a decade ago; the corresponding share for DM countries is much larger at 20%. This increased demand has in turn partly driven the current bull market in gold, with prices forecast to climb toward $4,000/oz by mid-2026. Bond markets In a sign of de-dollarization in bond markets, the share of foreign ownership in the U.S. Treasury market has been declining over the last 15 years. USD assets, principally liquid Treasuries, account for the majority of allocated FX reserves. However, demand for Treasuries has stagnated among foreign official institutions, as the growth of FX reserves has slowed and the USD’s share of reserves has dropped from its recent peak. Similarly, the backdrop for foreign private demand has weakened — as yields have risen across DM government bond markets, Treasuries have become relatively less attractive. While foreign investors remain the largest constituent within the Treasury market, their share of ownership has fallen to 30% as of early 2025 — down from a peak of above 50% during the GFC. “Although foreign demand has not kept pace with the growth of the Treasury market for more than a decade, we must consider what more aggressive action could mean. Japan is the largest foreign creditor and alone holds more than $1.1 trillion Treasuries, or nearly 4% of the market. Accordingly, any significant foreign selling would be impactful, driving yields higher,” said Jay Barry, head of Global Rates Strategy at J.P. Morgan. According to estimates by J.P. Morgan Research, each 1-percentage-point decline in foreign holdings relative to GDP (or approximately $300 billion of Treasuries) would result in yields rising by more than 33 basis points (bp). “While this is not our base case, it nonetheless underscores the impact of foreign investment on risk-free rates,” Barry added. Commodity markets De-dollarization is most visible in commodity markets, where the greenback’s influence on pricing has diminished. “Today, a large and growing proportion of energy is being priced in non-dollar-denominated contracts,” said Natasha Kaneva, head of Global Commodities Strategy at J.P. Morgan. For example, due to Western sanctions, Russian oil products exported eastward and southward are being sold in the local currencies of buyers, or in the currencies of countries Russia perceives as friendly. Among buyers, India, China and Turkey are all either using or seeking alternatives to the dollar. Saudi Arabia is also considering adding yuan-denominated futures contracts in the pricing model of Saudi Arabian oil, though progress has been slow. Notably, cross-border trade settlement in yuan is gaining ground outside of oil too. Some Indian companies have started paying for Russian coal imports in yuan, even without the involvement of Chinese intermediaries. Bangladesh also recently decided to pay Russia for its 1.4 GW nuclear power plant in yuan. “The de-dollarization trend in the commodity trade is a boon for countries like India, China, Brazil, Thailand and Indonesia, which can now not only buy oil at a discount, but also pay for it with their own local currencies,” Kaneva noted. “This reduces the need for precautionary reserves of U.S. dollars, U.S. Treasuries and oil, which might in turn free up capital to be deployed in growth-boosting domestic projects.” Deposit dollarization in emerging markets At the other end of the spectrum, deposit dollarization — where a significant portion of a country’s bank deposits are denominated in the U.S. dollar instead of the local currency — is still evident in many EM countries. “The tendency of EM residents to dollarize in times of stress appears to be correlated across markets,” said Jonny Goulden, head of EM Fixed Income Strategy at J.P. Morgan. According to J.P. Morgan Research, dollar deposits have grown mostly uninterrupted over the last decade in EM, reaching around $830 billion for a sample set of 18 EM countries (excluding China, Singapore and Hong Kong). “While there are large regional divergences in deposit dollarization across EM, all regions are more dollarized now than they were a decade ago,” Goulden noted. Latin America is the most dollarized region, with an aggregate dollarization rate of 19.1%. EMEA’s rate stands at 15.2%, while Asia (excluding China, Singapore and Hong Kong) has the lowest rate at 9.7%. China is the exception, as its dollarization rate has been persistently falling since 2017. “This is not surprising, as this was around the time when U.S.–China relations began shifting into their current state, marked by the trade war and growing diplomatic, security and geopolitical tensions,” Goulden said. “This suggests that China, alongside progress on de-dollarizing its own cross-border transactions, has effectively been de-dollarizing the deposits of Chinese residents, adding another dimension to its efforts to separate from U.S. dominance.”
考点1:“overhang in USD longs” 在这个语境里应该翻译成“过剩的美元多头头寸”或者“超额美元多头头寸” 考点2:“base case”推荐译为““基准情形”、“基本预测”或“基本假设” 考点3:“the share of foreign ownership in the U.S. Treasury market ” 中的share不能被翻译成“持股比例”
6223b
学术论文
社会科学
177
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 更让人困惑的是,有报道称神经干细胞(NSCs)可去分化为分化程度更低的祖细胞,甚至可分化为通常来源于其他两个胚层的细胞类型。 成熟生物体中的正常二倍体细胞含有几乎相同的DNA,而细胞类型是由特定基因的选择性转录或抑制所决定的。如今越来越明确的是,任何体细胞都有可能被重编程为分化程度更低的状态,尽管这一过程的潜在机制尚未明确。细胞融合研究已证实,多能干细胞中含有可重编程宿主体细胞核的因子,使其分化潜能的限制减少。例如,胸腺细胞与胚胎癌细胞融合后,可重新激活雄性来源胸腺细胞中的X染色体。同样,将室管膜下区(SVZ)神经干细胞与肌肉细胞共培养,可诱导其分化为表达特定肌肉标志物的肌细胞。 这种可塑性并非神经干细胞所特有,有大量报道显示非神经干细胞可转分化为神经前体形式。更有趣的是,从骨髓基质中分离出的间充质干细胞群可被诱导生成神经元、少突胶质细胞和星形胶质细胞。 也有证据表明体内存在生长因子介导的去分化现象。近藤(Kondo)和拉夫(Raff)通过骨形态发生蛋白(BMPs)处理纯化的少突胶质细胞祖细胞,使其成为神经干细胞。由此产生的细胞能够生成神经元和星形胶质细胞,不过尚未正式排除原始视神经制备物中存在微量污染干细胞的可能性。在另一项研究中,为帮助腿部溃疡愈合而接受人表皮生长因子(EGF)治疗的患者,其再生表皮中含有表达β-1整合素和角蛋白19的细胞,而这两种蛋白是表皮干细胞的标志物。然而,这些实验并未排除现有干细胞被募集到伤口部位的可能性。 不过,干细胞转分化的概念并未得到普遍认可。在一项研究中,比约森(Bjornson)等人将克隆来源的神经干细胞静脉注射到经辐射处理的小鼠尾静脉中,发现这些干细胞生成了大量造血细胞。但范德库伊(Van der Kooy)及其同事的研究未能重复上述结果——他们给128只宿主动物每只的尾静脉注射100万个神经球细胞,却未检测到来自标记神经球的大量造血细胞,研究认为早期实验中干细胞的长期培养可能是导致结果差异的原因。 最近,两项独立研究为转分化提供了另一种解释。第一项研究中,将绿色荧光蛋白(GFP)标记的神经元与潮霉素抗性胚胎干细胞(ES细胞)共培养,获得了具有神经元和ES细胞双重分子特征的未分化杂交细胞。第二项研究中,将GFP标记的骨髓细胞与ES细胞共培养,产生了GFP阳性的融合细胞,这些细胞表达多种ES细胞基因。值得注意的是,这些杂交细胞主要为四倍体,这表明骨髓细胞可与其他细胞类型自发融合并获得其表型。 当然,从伦理角度而言,转分化的争议备受关注。有观点认为,若成体干细胞可被诱导分化为任何细胞类型,那么对胚胎干细胞研究的需求就会减少。然而,大量研究仍表明,胎儿干细胞的可塑性显著高于成体干细胞。 神经发生与脑损伤 当前,许多研究聚焦于理解神经退行性疾病及脑损伤后的神经发生。这类研究既探究内源性修复的潜力,也旨在深入了解受损脑内的许可性和限制性信号——这对于未来考虑细胞替代疗法至关重要。现已明确,中枢神经系统(CNS)损伤确实会导致神经前体细胞增殖。 多项研究结合BrdU标记与神经元特异性标志物,证实了实验性卒中后内源性神经前体细胞的增殖与分化。例如,在沙鼠脑缺血模型中,研究人员利用BrdU结合多唾液酸神经细胞黏附分子(PSA-NCAM)和神经元核抗原(NeuN)分别标记神经干细胞和神经元,以探究神经干细胞的增殖、迁移和分化情况。损伤后20天,首先在齿状回颗粒下区检测到BrdU标记的神经干细胞,而NeuN阳性神经元的数量在损伤后60天内持续增加。 同样,沙鼠全脑缺血后,齿状回颗粒下区的神经发生增加了10倍。 卒中后的内源性修复还可能涉及室管膜下区(SVZ)神经祖细胞的增殖。大脑中动脉闭塞后,注射BrdU可特异性标记室管膜层和室管膜下层的细胞。有趣的是,这些细胞在卒中后立即被染色为星形胶质细胞,但在损伤后后期获得了神经元的特征性抗原标志物,这进一步支持了阿尔瓦雷斯-布伊拉(Alvarez-Buylla)的室管膜下区干细胞假说。在另一项采用化学诱导啮齿类动物癫痫发作的模型中,研究报道室管膜下区神经前体细胞的生成显著增加,且这些细胞随后向嗅球迁移并整合。BrdU标记证实了细胞增殖,逆转录病毒示踪证实了细胞迁移。尽管有人提出缺血诱导的神经发生可能有助于损伤后缺失的特定记忆功能恢复,但在损伤后的数周内,大部分分裂细胞会丢失。此类反应是特异性的,还是在已有成年神经发生的区域出现的一般性整体反应,仍有待证实。损伤后神经细胞类型的生成与存活现象,引发了人们对能否利用外源性细胞在细胞缺失的疾病和损伤中进行细胞替代的关注。 利用干细胞修复受损神经系统 如上所述,有令人振奋的证据表明,健康成年脑内存在神经细胞的增殖与分化,且损伤可进一步刺激这一过程。然而,神经细胞的再生能力有限,内源性神经干细胞数量较少,似乎无法在损伤后完全重建并恢复功能。这促使多个研究团队探究脑损伤后细胞替代疗法的潜力。 这些研究充分表明,在适宜条件下,细胞移植物可在脑内整合并发挥功能。 细胞替代的主要来源有三种:胎儿神经组织、永生化细胞系和胚胎干细胞(ES细胞)。胎儿神经细胞移植已用于多种脑损伤模型。在成年大鼠局灶性前脑缺血损伤后,胎儿皮质移植物可在梗死区域存活,且似乎能接收来自周围脑组织的连接,进而改善运动功能、空间学习和记忆能力。在患有神经退行性疾病的人类中,胎儿神经细胞移植已有应用经验。目前,已有300多名帕金森病患者接受了移植到纹状体的胎儿中脑前体细胞。这些移植物具有自主活性,可将多巴胺释放恢复至接近正常水平,并改善症状。然而,存在一个弊端:在丹佛和纽约的近期临床试验中,15%的移植患者出现了难以接受的运动障碍。此外,胎儿神经组织的供应有限,因此可获得的神经元数量较少。这一问题可通过体外扩增部分解决。令人鼓舞的是,培养扩增的胚胎12天神经干细胞仍保留分化为多巴胺能神经元的能力,并能改善帕金森病大鼠模型的结局。移植物的免疫排斥及后续失败也是一个问题,尤其是在胎儿多巴胺能神经元异种移植中,移植物存活率低且无临床改善。胚胎细胞的一个潜在优势是可通过克隆技术进行修饰,以表达患者自身的基因型。该技术可用于提供免疫相容性细胞用于移植。然而,有观点认为,ES细胞在培养中无限增殖的能力并非导致移植排斥,而是给受者带来了肿瘤发生的实际风险。因此,使用成体干细胞而非胚胎来源干细胞可能不仅出于伦理原因。 作为胎儿组织的替代物,永生化细胞系已用于脑损伤的动物研究。将人畸胎瘤细胞系来源的神经元移植到局灶性缺血大鼠体内,实现了组织学整合并改善了功能;将海马神经上皮细胞系移植到小鼠受损海马中也得到了类似结果。多项研究还证实,少突胶质细胞祖细胞(OPCs)可成功移植用于治疗脱髓鞘疾病。在通过注射髓鞘碱性蛋白诱导实验性自身免疫性脑脊髓炎(EAE)前,将OPCs移植到大鼠脊髓中,其可长期存活。令人印象深刻的是,所用细胞系(CG4)在健康大鼠中可迁移至6厘米远的神经纤维束,且尽管经过转化,仍未出现失控增殖。移植的祖细胞与反应性星形胶质细胞的密切联系表明,与当前的传统观点相反,受损中枢神经系统内的炎症信号实际上可促进整合。这些结果与富兰克林(Franklin)等人的研究一致,他们发现CG4细胞也可整合到溴化乙锭诱导的脊髓脱髓鞘病灶中。在该研究中,细胞同样从注射部位迁移至脱髓鞘区域重新定植,且再次无法穿透未受损的白质组织。然而,人们严重担忧永生化细胞系易发生肿瘤发生,且无法重建脑损伤中丢失的多种细胞类型。这使得它们在临床应用中仅具有有限价值。
更让人困惑的是,有报道称神经干细胞(NSCs)可去分化为分化程度更低的祖细胞,甚至可分化为通常来源于其他两个胚层的细胞类型。 成熟生物体中的正常二倍体细胞含有几乎相同的DNA,而细胞类型是由特定基因的选择性转录或抑制所决定的。如今越来越明确的是,任何体细胞都有可能被重编程为分化程度更低的状态,尽管这一过程的潜在机制尚未明确。细胞融合研究已证实,多能干细胞中含有可重编程宿主体细胞核的因子,使其分化潜能的限制减少。例如,胸腺细胞与胚胎癌细胞融合后,可重新激活雄性来源胸腺细胞中的X染色体。同样,将室管膜下区(SVZ)神经干细胞与肌肉细胞共培养,可诱导其分化为表达特定肌肉标志物的肌细胞。 这种可塑性并非神经干细胞所特有,有大量报道显示非神经干细胞可转分化为神经前体形式。更有趣的是,从骨髓基质中分离出的间充质干细胞群可被诱导生成神经元、少突胶质细胞和星形胶质细胞。 也有证据表明体内存在生长因子介导的去分化现象。近藤(Kondo)和拉夫(Raff)通过骨形态发生蛋白(BMPs)处理纯化的少突胶质细胞祖细胞,使其成为神经干细胞。由此产生的细胞能够生成神经元和星形胶质细胞,不过尚未正式排除原始视神经制备物中存在微量污染干细胞的可能性。在另一项研究中,为帮助腿部溃疡愈合而接受人表皮生长因子(EGF)治疗的患者,其再生表皮中含有表达β-1整合素和角蛋白19的细胞,而这两种蛋白是表皮干细胞的标志物。然而,这些实验并未排除现有干细胞被募集到伤口部位的可能性。 不过,干细胞转分化的概念并未得到普遍认可。在一项研究中,比约森(Bjornson)等人将克隆来源的神经干细胞静脉注射到经辐射处理的小鼠尾静脉中,发现这些干细胞生成了大量造血细胞。但范德库伊(Van der Kooy)及其同事的研究未能重复上述结果——他们给128只宿主动物每只的尾静脉注射100万个神经球细胞,却未检测到来自标记神经球的大量造血细胞,研究认为早期实验中干细胞的长期培养可能是导致结果差异的原因。 最近,两项独立研究为转分化提供了另一种解释。第一项研究中,将绿色荧光蛋白(GFP)标记的神经元与潮霉素抗性胚胎干细胞(ES细胞)共培养,获得了具有神经元和ES细胞双重分子特征的未分化杂交细胞。第二项研究中,将GFP标记的骨髓细胞与ES细胞共培养,产生了GFP阳性的融合细胞,这些细胞表达多种ES细胞基因。值得注意的是,这些杂交细胞主要为四倍体,这表明骨髓细胞可与其他细胞类型自发融合并获得其表型。 当然,从伦理角度而言,转分化的争议备受关注。有观点认为,若成体干细胞可被诱导分化为任何细胞类型,那么对胚胎干细胞研究的需求就会减少。然而,大量研究仍表明,胎儿干细胞的可塑性显著高于成体干细胞。 神经发生与脑损伤 当前,许多研究聚焦于理解神经退行性疾病及脑损伤后的神经发生。这类研究既探究内源性修复的潜力,也旨在深入了解受损脑内的许可性和限制性信号——这对于未来考虑细胞替代疗法至关重要。现已明确,中枢神经系统(CNS)损伤确实会导致神经前体细胞增殖。 多项研究结合BrdU标记与神经元特异性标志物,证实了实验性卒中后内源性神经前体细胞的增殖与分化。例如,在沙鼠脑缺血模型中,研究人员利用BrdU结合多唾液酸神经细胞黏附分子(PSA-NCAM)和神经元核抗原(NeuN)分别标记神经干细胞和神经元,以探究神经干细胞的增殖、迁移和分化情况。损伤后20天,首先在齿状回颗粒下区检测到BrdU标记的神经干细胞,而NeuN阳性神经元的数量在损伤后60天内持续增加。 同样,沙鼠全脑缺血后,齿状回颗粒下区的神经发生增加了10倍。 卒中后的内源性修复还可能涉及室管膜下区(SVZ)神经祖细胞的增殖。大脑中动脉闭塞后,注射BrdU可特异性标记室管膜层和室管膜下层的细胞。有趣的是,这些细胞在卒中后立即被染色为星形胶质细胞,但在损伤后后期获得了神经元的特征性抗原标志物,这进一步支持了阿尔瓦雷斯-布伊拉(Alvarez-Buylla)的室管膜下区干细胞假说。在另一项采用化学诱导啮齿类动物癫痫发作的模型中,研究报道室管膜下区神经前体细胞的生成显著增加,且这些细胞随后向嗅球迁移并整合。BrdU标记证实了细胞增殖,逆转录病毒示踪证实了细胞迁移。尽管有人提出缺血诱导的神经发生可能有助于损伤后缺失的特定记忆功能恢复,但在损伤后的数周内,大部分分裂细胞会丢失。此类反应是特异性的,还是在已有成年神经发生的区域出现的一般性整体反应,仍有待证实。损伤后神经细胞类型的生成与存活现象,引发了人们对能否利用外源性细胞在细胞缺失的疾病和损伤中进行细胞替代的关注。 利用干细胞修复受损神经系统 如上所述,有令人振奋的证据表明,健康成年脑内存在神经细胞的增殖与分化,且损伤可进一步刺激这一过程。然而,神经细胞的再生能力有限,内源性神经干细胞数量较少,似乎无法在损伤后完全重建并恢复功能。这促使多个研究团队探究脑损伤后细胞替代疗法的潜力。 这些研究充分表明,在适宜条件下,细胞移植物可在脑内整合并发挥功能。 细胞替代的主要来源有三种:胎儿神经组织、永生化细胞系和胚胎干细胞(ES细胞)。胎儿神经细胞移植已用于多种脑损伤模型。在成年大鼠局灶性前脑缺血损伤后,胎儿皮质移植物可在梗死区域存活,且似乎能接收来自周围脑组织的连接,进而改善运动功能、空间学习和记忆能力。在患有神经退行性疾病的人类中,胎儿神经细胞移植已有应用经验。目前,已有300多名帕金森病患者接受了移植到纹状体的胎儿中脑前体细胞。这些移植物具有自主活性,可将多巴胺释放恢复至接近正常水平,并改善症状。然而,存在一个弊端:在丹佛和纽约的近期临床试验中,15%的移植患者出现了难以接受的运动障碍。此外,胎儿神经组织的供应有限,因此可获得的神经元数量较少。这一问题可通过体外扩增部分解决。令人鼓舞的是,培养扩增的胚胎12天神经干细胞仍保留分化为多巴胺能神经元的能力,并能改善帕金森病大鼠模型的结局。移植物的免疫排斥及后续失败也是一个问题,尤其是在胎儿多巴胺能神经元异种移植中,移植物存活率低且无临床改善。胚胎细胞的一个潜在优势是可通过克隆技术进行修饰,以表达患者自身的基因型。该技术可用于提供免疫相容性细胞用于移植。然而,有观点认为,ES细胞在培养中无限增殖的能力并非导致移植排斥,而是给受者带来了肿瘤发生的实际风险。因此,使用成体干细胞而非胚胎来源干细胞可能不仅出于伦理原因。 作为胎儿组织的替代物,永生化细胞系已用于脑损伤的动物研究。将人畸胎瘤细胞系来源的神经元移植到局灶性缺血大鼠体内,实现了组织学整合并改善了功能;将海马神经上皮细胞系移植到小鼠受损海马中也得到了类似结果。多项研究还证实,少突胶质细胞祖细胞(OPCs)可成功移植用于治疗脱髓鞘疾病。在通过注射髓鞘碱性蛋白诱导实验性自身免疫性脑脊髓炎(EAE)前,将OPCs移植到大鼠脊髓中,其可长期存活。令人印象深刻的是,所用细胞系(CG4)在健康大鼠中可迁移至6厘米远的神经纤维束,且尽管经过转化,仍未出现失控增殖。移植的祖细胞与反应性星形胶质细胞的密切联系表明,与当前的传统观点相反,受损中枢神经系统内的炎症信号实际上可促进整合。这些结果与富兰克林(Franklin)等人的研究一致,他们发现CG4细胞也可整合到溴化乙锭诱导的脊髓脱髓鞘病灶中。在该研究中,细胞同样从注射部位迁移至脱髓鞘区域重新定植,且再次无法穿透未受损的白质组织。然而,人们严重担忧永生化细胞系易发生肿瘤发生,且无法重建脑损伤中丢失的多种细胞类型。这使得它们在临床应用中仅具有有限价值。
考点1:“间充质干细胞” 只能翻译为 “mesenchymal stem cells”,因为这是医学术语 考点2:“骨形态发生蛋白(BMPs)” 只能翻译为 “bone morphogenetic proteins (BMPs)”,因为这是医学术语 考点3:“绿色荧光蛋白(GFP)” 只能翻译为 “green fluorescent protein (GFP)”,因为这是医学术语 考点4:“少突胶质细胞祖细胞(OPCs)” 只能翻译为 “oligodendrocyte progenitor cells (OPCs)”,因为这是医学术语 考点5:“多唾液酸神经细胞黏附分子(PSA-NCAM)” 只能翻译为 “polysialylated neural cell adhesion molecule (PSA-NCAM)”,因为这是医学术语 考点6:“神经元核抗原(NeuN)” 只能翻译为 “neuronal nuclear antigen (NeuN)”,因为这是医学术语 考点7:“大脑中动脉闭塞” 只能翻译为 middle cerebral artery occlusion”,因为这是医学术语 考点8:“多巴胺能神经元” 只能翻译为 “dopaminergic neurones”,因为这是医学术语 考点9:“畸胎瘤” 只能翻译为 “teratocarcinoma”,因为这是医学术语 考点10:“全脑缺血” 只能翻译为 “global cerebral ischemia ”,因为这是医学术语 考点11:“运动障碍” 只能翻译为 “dyskinesias”,强调患者是病理性不自主运动。不可翻译为 “motor disturbances”, 考点12:“传统观点” 在文中指神经科学领域中长期公认的学术观点,在科学文献中常用 “dogma” 表示这类“被普遍接受的教条式观点”。需要翻译为 “current dogma”。不能翻译为 “current conventional views”,“conventional views” 仅指“传统看法”,无法体现文中语境表现的“学术领域内长期公认的教条式观点”这一深层含义。 考点13:“异种移植” 只能翻译为“xenotransplantation”。 考点14:“前神经体细胞” 只能翻译为 “neural progenitor cell”,因为这是医学术语
62df9
学术论文
自然科学
197
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Summary Companies want to rapidly digitize procurement by exploiting emerging technologies like AI. However, traditional IT investment and adoption processes are often too slow and rigid. To overcome the obstacles, BP, Otto Group, and Walmart have developed new approaches. BP’s procurement technology garage conducts rapid pilots to explore emerging technologies and assess their ROI potential. Otto Group’s new-technology ventures team engages with startups to experiment and co-develop solutions. Walmart’s procurement center of excellence scales proven technologies globally and maximizes their ROI.close Companies have ambitious plans for rapidly digitizing the procurement function, especially through emerging technologies like AI. That was a finding of a study that we conducted with Digital Procurement World in October 2024, which included a survey of over 200 procurement executives around the world. However, in separate in-depth interviews we conducted with over two dozen procurement executives and a few of their colleagues in the supply chain and IT functions, we discovered obstacles that can get in the way of these ambitions. But we also found ways that three companies featured in this article are successfully employing to overcome them: BP has a procurement technology garage to quickly explore emerging technologies from multiple startups simultaneously and to develop a fact-based understanding of their ROI potential. Otto Group, an e-commerce and retail company based in Hamburg, Germany, whose businesses include Crate and Barrel and Bonprix, has a new-technology ventures team to engage with startups and emerging technologies to accelerate experimentation and co-develop applications. Walmart has a procurement technology center of excellence whose mission is to roll out proven technologies globally at scale and maximize their ROI. The Need for New Approaches One big obstacle we discovered is the traditional IT investment and adoption process. “Traditional contract negotiation processes managed by IT are not geared for startup environments,” Arno Baltruschat, head of Otto’s new-technology ventures team, told us. Conventional approaches—often centered on large vendors and marked by lengthy negotiation and contracting cycles—are ill-suited for exploring promising digital innovations that startups are developing. Companies historically use rigid contract templates that are more than 50 pages long, and it often takes months to finalize them. That is overkill and much too slow when dealing with startups. The rigidity of these contracts, which typically involve a multi-year commitment to a technology based upon anticipated benefits and a predetermined rollout plan, limits the innovation potential of the relationships and ignores the need to experiment to explore the full potential of the new technologies. Another challenge is scaling a successful pilot project that focused on a handful of use cases for a new technology to the broader global organization. Procurement functions often struggle to do this. Now let’s look at BP’s, Otto’s, and Walmart’s remedies. BP’s Procurement Technology Garage This BP unit conducts rapid small-scale pilots to explore emerging technologies that might provide a solution for an existing procurement challenge. In doing so, the garage also develops a fact-based understanding of the technology’s ROI potential (or lack thereof). After a pilot has been concluded, the garage team offers a recommendation to the chief procurement officer that can range from “do not recommend further evaluation” to “recommend full-scale adoption” (with the results of the pilot, the insights it generated, and the business case). When a pilot does not deliver promising results, it is seen as a learning experience, not a failure. Intermediate options include refining the pilot and collaborating with the technology provider to co-develop additional features. The garage has been in existence since late 2023 and has conducted pilots in a variety of technology areas, from sustainable procurement to AI-supported contractor invoicing to procurement strategy development. Pilots typically last between three to six months. Three technologies have already been given the green light for either further development or full-scale adoption. How the Garage Works Before launching a pilot, the garage team identifies significant procurement challenges or opportunities, then seeks out potential technology solutions to address them. Nicholas Wright, head of BP’s garage, states: “Our ethos is problem first, solution search after. We are data-led in the selection of problems.” For example, the team found that procurement managers were spending weeks, sometimes months, on the development of category strategies for their categories of spend. (A category strategy captures the organization’s approach for getting the best out of the supplier market given its business needs and supply market characteristics and provides the basis for tendering, contracting, and managing supplier relationships.) The garage team found that upon completion, strategies would often remain unused until it was time to update them. “The strategies were not realizing their full potential and were not ‘lived’ in the organization,” Wright said. The garage team scanned the technology landscape and found a promising AI-enabled tool called Cirtuo that supports the development of category strategies and proposed to the procurement leaders that it be allowed to pilot this technology for a three-month period. They agreed and nominated categories and buyer teams from across the business for inclusion in the pilot. During the pilot, the buyer teams used the tool to develop new strategies for their categories. The strategies created during the pilot were of significantly higher quality than those developed without the tool, featuring, on average, 10 times more insights. The strategies were also more multi-dimensional, addressing all areas of potential value creation—not only cost savings but also company priorities like sustainability. The AI features of the tool that support the collection and analysis of external supplier market data not only enriched strategies but also reduced development time by 50%, on average. Crucially, the time savings have allowed category managers to spend less time on research and strategy documentation and more on driving organizational engagement with the strategies, reducing the risk of them being shelved and unused. Based upon the evaluation report of the garage team, BP’s chief procurement officer gave the green light for scaling the tool across the global procurement organization. Otto’s Ventures Client Unit Otto’s unit, called OTTO DOCK 6, offers an approach for engaging with startups and emerging technologies to experiment and possibly even co-develop technology. Unlike traditional internal venture capital groups, this group does not invest financially in startups; instead it co-develops and pilots applications with startup solutions to pressing business challenges before handing the technology over to the business. Arno Baltruschat, co-leader of the unit, told us: “Startups can react fast to our needs and we can move faster to partner with startups because we don’t need the due diligence check and overhead required for a Series A investment by a venture capital firm.” In the last four years, the venture client unit established non-investment partnerships with 50 startups. How the Venture Client Unit Works The unit’s approach begins with a sizable, strategically important business problem, not with a promising startup or interesting technology. Once high-value problems are identified, the unit scouts startups and pilots and may co-develop an application of the technology in collaboration with the startup and business colleagues to use the technology in Otto operations. The contracts for these engagements are three- to six-page agreements that typically take no more than a few weeks to finalize. Given that the aim of pilots is rapid learning and development, they are intentionally limited in scale, scope, and duration (two to three months). After a pilot has ended, the group evaluates results and may recommend a further rollout if substantial ROI hurdles are likely to be met. Currently, only about 40% of technologies are adopted. Successful projects with tech startups include automated product damage checks and ESG scoring of products. AI-based automated negotiations became a recent focus of the venture client unit. In 2023, it began seeking tech startups to help more efficiently manage long-tail suppliers, the large number of companies with low-value agreements to provide non-strategic things. It selected Pactum, a startup specializing in AI-powered negotiations software. In early 2024, the unit partnered with Pactum and collaborated with leaders and buyers from Otto’s corporate procurement department and supply-chain-management function to test whether the AI chatbot could negotiate simple terms like payment days and discounts with suppliers. Buyer involvement was essential to the success of the pilot, particularly in communicating with suppliers and selecting suitable participants. Buyers played a key role in identifying spend categories and suppliers that would ensure the pilot was minimally disruptive to core business operations. To reduce operational risk, they focused on less-strategic products and services. To limit financial exposure, they selected suppliers with annual spend levels between €100,000 and €250,000. In total, the pilot encompassed approximately €24 million in spending across all qualifying suppliers. Of the suppliers invited to participate, nearly 65% accepted. Among those who engaged in the pilot, 70% successfully closed agreements, generating incremental savings for Otto. Based on the success of the first pilot, the team next invited larger suppliers, involving greater amounts of external spend. This first phase was successfully completed in just three months. By the end of 2024, Otto presented a strategic implementation plan to senior management that proposed expanding automated negotiations to more complex categories and to more intricate terms, including pricing and bonuses. This year, the project will transition from the venture client unit to an operational excellence center responsible for rolling out digital technologies throughout the procurement organization and further enhancing the technology for new applications.
Summary Companies want to rapidly digitize procurement by exploiting emerging technologies like AI. However, traditional IT investment and adoption processes are often too slow and rigid. To overcome the obstacles, BP, Otto Group, and Walmart have developed new approaches. BP’s procurement technology garage conducts rapid pilots to explore emerging technologies and assess their ROI potential. Otto Group’s new-technology ventures team engages with startups to experiment and co-develop solutions. Walmart’s procurement center of excellence scales proven technologies globally and maximizes their ROI.close Companies have ambitious plans for rapidly digitizing the procurement function, especially through emerging technologies like AI. That was a finding of a study that we conducted with Digital Procurement World in October 2024, which included a survey of over 200 procurement executives around the world. However, in separate in-depth interviews we conducted with over two dozen procurement executives and a few of their colleagues in the supply chain and IT functions, we discovered obstacles that can get in the way of these ambitions. But we also found ways that three companies featured in this article are successfully employing to overcome them: BP has a procurement technology garage to quickly explore emerging technologies from multiple startups simultaneously and to develop a fact-based understanding of their ROI potential. Otto Group, an e-commerce and retail company based in Hamburg, Germany, whose businesses include Crate and Barrel and Bonprix, has a new-technology ventures team to engage with startups and emerging technologies to accelerate experimentation and co-develop applications. Walmart has a procurement technology center of excellence whose mission is to roll out proven technologies globally at scale and maximize their ROI. The Need for New Approaches One big obstacle we discovered is the traditional IT investment and adoption process. “Traditional contract negotiation processes managed by IT are not geared for startup environments,” Arno Baltruschat, head of Otto’s new-technology ventures team, told us. Conventional approaches—often centered on large vendors and marked by lengthy negotiation and contracting cycles—are ill-suited for exploring promising digital innovations that startups are developing. Companies historically use rigid contract templates that are more than 50 pages long, and it often takes months to finalize them. That is overkill and much too slow when dealing with startups. The rigidity of these contracts, which typically involve a multi-year commitment to a technology based upon anticipated benefits and a predetermined rollout plan, limits the innovation potential of the relationships and ignores the need to experiment to explore the full potential of the new technologies. Another challenge is scaling a successful pilot project that focused on a handful of use cases for a new technology to the broader global organization. Procurement functions often struggle to do this. Now let’s look at BP’s, Otto’s, and Walmart’s remedies. BP’s Procurement Technology Garage This BP unit conducts rapid small-scale pilots to explore emerging technologies that might provide a solution for an existing procurement challenge. In doing so, the garage also develops a fact-based understanding of the technology’s ROI potential (or lack thereof). After a pilot has been concluded, the garage team offers a recommendation to the chief procurement officer that can range from “do not recommend further evaluation” to “recommend full-scale adoption” (with the results of the pilot, the insights it generated, and the business case). When a pilot does not deliver promising results, it is seen as a learning experience, not a failure. Intermediate options include refining the pilot and collaborating with the technology provider to co-develop additional features. The garage has been in existence since late 2023 and has conducted pilots in a variety of technology areas, from sustainable procurement to AI-supported contractor invoicing to procurement strategy development. Pilots typically last between three to six months. Three technologies have already been given the green light for either further development or full-scale adoption. How the Garage Works Before launching a pilot, the garage team identifies significant procurement challenges or opportunities, then seeks out potential technology solutions to address them. Nicholas Wright, head of BP’s garage, states: “Our ethos is problem first, solution search after. We are data-led in the selection of problems.” For example, the team found that procurement managers were spending weeks, sometimes months, on the development of category strategies for their categories of spend. (A category strategy captures the organization’s approach for getting the best out of the supplier market given its business needs and supply market characteristics and provides the basis for tendering, contracting, and managing supplier relationships.) The garage team found that upon completion, strategies would often remain unused until it was time to update them. “The strategies were not realizing their full potential and were not ‘lived’ in the organization,” Wright said. The garage team scanned the technology landscape and found a promising AI-enabled tool called Cirtuo that supports the development of category strategies and proposed to the procurement leaders that it be allowed to pilot this technology for a three-month period. They agreed and nominated categories and buyer teams from across the business for inclusion in the pilot. During the pilot, the buyer teams used the tool to develop new strategies for their categories. The strategies created during the pilot were of significantly higher quality than those developed without the tool, featuring, on average, 10 times more insights. The strategies were also more multi-dimensional, addressing all areas of potential value creation—not only cost savings but also company priorities like sustainability. The AI features of the tool that support the collection and analysis of external supplier market data not only enriched strategies but also reduced development time by 50%, on average. Crucially, the time savings have allowed category managers to spend less time on research and strategy documentation and more on driving organizational engagement with the strategies, reducing the risk of them being shelved and unused. Based upon the evaluation report of the garage team, BP’s chief procurement officer gave the green light for scaling the tool across the global procurement organization. Otto’s Ventures Client Unit Otto’s unit, called OTTO DOCK 6, offers an approach for engaging with startups and emerging technologies to experiment and possibly even co-develop technology. Unlike traditional internal venture capital groups, this group does not invest financially in startups; instead it co-develops and pilots applications with startup solutions to pressing business challenges before handing the technology over to the business. Arno Baltruschat, co-leader of the unit, told us: “Startups can react fast to our needs and we can move faster to partner with startups because we don’t need the due diligence check and overhead required for a Series A investment by a venture capital firm.” In the last four years, the venture client unit established non-investment partnerships with 50 startups. How the Venture Client Unit Works The unit’s approach begins with a sizable, strategically important business problem, not with a promising startup or interesting technology. Once high-value problems are identified, the unit scouts startups and pilots and may co-develop an application of the technology in collaboration with the startup and business colleagues to use the technology in Otto operations. The contracts for these engagements are three- to six-page agreements that typically take no more than a few weeks to finalize. Given that the aim of pilots is rapid learning and development, they are intentionally limited in scale, scope, and duration (two to three months). After a pilot has ended, the group evaluates results and may recommend a further rollout if substantial ROI hurdles are likely to be met. Currently, only about 40% of technologies are adopted. Successful projects with tech startups include automated product damage checks and ESG scoring of products. AI-based automated negotiations became a recent focus of the venture client unit. In 2023, it began seeking tech startups to help more efficiently manage long-tail suppliers, the large number of companies with low-value agreements to provide non-strategic things. It selected Pactum, a startup specializing in AI-powered negotiations software. In early 2024, the unit partnered with Pactum and collaborated with leaders and buyers from Otto’s corporate procurement department and supply-chain-management function to test whether the AI chatbot could negotiate simple terms like payment days and discounts with suppliers. Buyer involvement was essential to the success of the pilot, particularly in communicating with suppliers and selecting suitable participants. Buyers played a key role in identifying spend categories and suppliers that would ensure the pilot was minimally disruptive to core business operations. To reduce operational risk, they focused on less-strategic products and services. To limit financial exposure, they selected suppliers with annual spend levels between €100,000 and €250,000. In total, the pilot encompassed approximately €24 million in spending across all qualifying suppliers. Of the suppliers invited to participate, nearly 65% accepted. Among those who engaged in the pilot, 70% successfully closed agreements, generating incremental savings for Otto. Based on the success of the first pilot, the team next invited larger suppliers, involving greater amounts of external spend. This first phase was successfully completed in just three months. By the end of 2024, Otto presented a strategic implementation plan to senior management that proposed expanding automated negotiations to more complex categories and to more intricate terms, including pricing and bonuses. This year, the project will transition from the venture client unit to an operational excellence center responsible for rolling out digital technologies throughout the procurement organization and further enhancing the technology for new applications.
考点 1:【procurement technology garage】应译为 【采购技术车库】 考点 2:【new-technology ventures team】 应译为 【新技术创投团队】 考点 3:【tendering, contracting, and managing supplier relationships】应译为 【招标、签约及供应商关系管理】 考点 4:【AI-enabled tool】应译为 【AI赋能工具】 考点 5:【venture client unit】应译为 【创投客户单元】 考点 6:【AI-based automated negotiations】应译为【基于AI的自动化谈判】 考点 7:【long-tail suppliers】应译为 【长尾供应商】 考点 8:【AI-powered negotiations software】应译为 【AI驱动谈判软件 / AI赋能谈判软件】 考点 9:【spend categories】应译为 【支出品类】 考点 10:【less-strategic products and services】应译为 【非战略性产品和服务】
65bdb
新闻资讯
观点评论
88
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: What is the keto diet, and can it be beneficial for you? You may have heard of friends and family trying the keto diet. It's a strict, high-fat diet with varying levels of protein and generally very low levels of carbohydrate. The keto diet was first used as a treatment for pediatric drug-resistant epilepsy to reduce seizures in the 1920s. It's still used for help prevent seizures in the U.S. when medications alone aren't enough. However, it can also help you lose weight if you stick with the plan long term. It may also provide other health benefits that come with weight loss. If you're wondering if the keto diet is right for you, our registered dietitians explain the eating plan and its pros and cons. What is the keto diet? The ketogenic or keto diet includes eating high-fat, low-carbohydrate foods to reach ketosis. In ketosis, the body uses fat for fuel, instead of sugar (glucose). This produces acids known as ketones. What are the basic rules of the keto diet? There are several different styles of keto diets. Each calls for eating different amounts of fat, protein and carbs in a day. Classic Keto Diet This is the original and best-studied of the different diet strategies. It's recommended that you work with a registered dietitian as it's the strictest, most difficult diet to follow. This diet is very in high fat, very low in carbs and relatively low in protein. All foods must be weighed on a scale to achieve a relative weight distribution of 4 grams of fat for every 1 gram of combined protein and carb eaten throughout the day. In other words, you will likely need to avoid all carb-rich foods (like pasta, potatoes, fruit, etc.). Also, for every serving of protein-rich food, you need four servings of a fatty food (heavy cream, avocado, olive oil, full fat cheese, etc.). A typical meal might look like half an avocado with a small piece of salmon cooked in a lot of olive oil with a side salad. It will all be topped with a cream sauce made with heavy cream and cheese. You would need to eat all of the oils and creams. Modified Keto Diet This is a modified form of the classic keto diet, allowing slightly more protein and carbs. This diet generally still requires a scale for weighing food. It's defined by a ratio closer to 3 grams of fat for every 1 gram of combined protein and carb (generally with a focus on more protein than carbs). In other words, you'll still avoid most carb-rich foods except for maybe one or two servings per day. For every serving of a protein- or carb-rich food, you'll need three servings of very high-fat food. Modified Atkins Diet This diet is the more user-friendly keto diet and likely the most popular. It's a high-fat, low-carb diet where you can choose to eat as much protein as you like. What foods can you eat on the keto diet? Here's a list of foods you should eat on the keto diet: Dairy and unsweetened dairy alternatives such as full fat cheese, plain full fat yogurt, cottage cheese and unsweetened plant-based milk. Low-carb vegetables like leafy greens, peppers and broccoli. Other plant-based foods including nuts, seeds and avocados. Oils and fats like olive and coconut oils, as well as whole olives and avocados. Moderate amounts of high protein foods such as seafood, meat, poultry and eggs. Sometimes small amounts of berries can be eaten. These diets generally avoid all starchy foods (pasta, rice, potatoes, bread, corn, peas, beans and legumes). Desserts and any other high-carb foods are also not part of the diet. Fruits can sometimes be included if they are high in fiber, such as raspberries. What are the pros and cons of the keto diet? Here are some benefits of the keto diet for your health: Helps with weight loss: Research suggests that the keto diet helps people cut calories because the diet doesn't include many foods that people are likely to overeat. However, it can be difficult to transition off of the keto diet without regaining weight. That's because people typically start eating more high-carb foods. Some data supports going from the keto diet to the Mediterranean diet or MIND diet may help maintain healthier weight once weight loss goals are achieved. Benefits people with type 2 diabetes: Some people with diabetes find that reducing their carb intake can help with blood sugar and medication management. Epilepsy management: Data shows that the keto diet can have positive effects for people with epilepsy in helping with seizure management. Some cons of the keto diet include: Can lead to nutrient deficiencies: The keto diet is very low in carbs which means you can't eat many fruits and vegetables. That can lead to low levels of vitamins A, E and B6, as well as folate, calcium, magnesium, and potassium. Possible negative effects on heart health: There are mixed results in studies that examine the specific effects of the keto diet on heart health. This is likely because much of it depends on individual food choices. Generally, foods high in saturated fat and a high-fat diet have negative consequences on heart health factors such as your lipid panel. Constipation and other gastrointestinal (GI) troubles: Constipation is likely the greatest long-term complication of the keto diet because of a low fiber intake. Fiber is a form of carb that is digested not by our human cells, but by the microbes in our gut. Any high-fiber foods (like beans, whole grains, and certain fruits) are also high in carbs, which is not typically part of the keto diet. An ongoing diet that is low in fiber (including the typical Western diet) may leave you at higher risk for certain cancers like colon cancer. What are some side effects of the keto diet? Before starting the keto diet, you should get a blood test to look at your lipids. You should then take another blood test after three to six months on the keto diet. Some people will actually find that their high-density lipoproteins (HDL) – the good cholesterol – go up and their triglycerides go down. But some people won't see a positive shift in those markers and may find their low-density lipoproteins (LDL) – the bad cholesterol – go up to unhealthy levels. There can be issues with bone health when people switch to low-carb keto diets. Researchers see that markers for bone breakdown are higher and markers for bone building are lower. This is a common side effect in kids who are fed the keto diet to treat epilepsy. Also, when people start the keto diet, they go through the keto flu. Keto flu symptoms appear a few days after starting the diet. This eventually goes away once your body adjusts to your eating changes. People on the keto diet also tend to lose more salt/sodium than those on a higher carb diet and may need more salt than the average person. Most people following a strict keto diet should take a multivitamin with minerals to ensure they're meeting those vitamin and mineral needs. The diet itself generally is not enough. Does the keto diet improve brain functions, such as learning and memory? The keto diet can show positive results for helping the brain. It's been used for years to treat epilepsy. Researchers are looking to see if the keto diet can help treat age-related brain decline and potentially Alzheimer's disease. There are mixed results on if the keto diet can help those with neuromuscular diseases such as Parkinson's disease. There are many different theories as to why the keto diet may be beneficial for brain health. These include the direct benefit from the presence of ketone bodies, especially one called beta hydroxybutyrate. Other theories suggest a possible benefit in changing the microbiome for certain conditions, like epilepsy. Is the keto diet good for athletes? Our bodies use multiple energy systems to exercise. The type of energy system used depends on the exercise intensity and duration. At lower intensities, our bodies use a mix of both burning fat and carbs for fuel. When we transition to high intensities (such as a scheduled workout), our bodies start to rely primarily on carbs. Most studies involving endurance athletes showed the keto diet hurts their performance. The keto diet has also been shown to increase losses of lean tissue when compared to a higher-carb diet. The bottom line is that the keto diet impairs athletic performance and are not recommended for athletes because they restrict carbs. What other eating plans might be right for me? If the keto diet doesn't sound like the right fit for you, you may want to consider one of these other eating plans. Intermittent fasting This weight-loss trend dates back to ancient times where fasting helps heal the body. The idea behind intermittent fasting is that by restricting food, our bodies will more quickly tap into fat stores for energy. It's not necessarily an eating plan that outlines what types of foods to eat but instead focuses on when to eat. There are a couple of different ways to implement intermittent fasting. Mediterranean diet This way of eating focuses on whole, plant-based foods and healthy fat. The Mediterranean diet's methods are scientifically proven to reduce the risk of chronic conditions like heart disease. MIND Diet Short for "Mediterranean-DASH Intervention for Neurodegenerative Delay," this diet is similar to the Mediterranean diet, but with stricter requirements for what you can eat. It emphasizes eating more vegetables, fruits (specifically berries), high-fiber foods, beans, nuts, seafood, poultry and olive oil. The MIND diet can possibly help reduce the cognitive decline associated with aging. It also may help slow disease progression for conditions like Parkinson's disease.
What is the keto diet, and can it be beneficial for you? You may have heard of friends and family trying the keto diet. It's a strict, high-fat diet with varying levels of protein and generally very low levels of carbohydrate. The keto diet was first used as a treatment for pediatric drug-resistant epilepsy to reduce seizures in the 1920s. It's still used for help prevent seizures in the U.S. when medications alone aren't enough. However, it can also help you lose weight if you stick with the plan long term. It may also provide other health benefits that come with weight loss. If you're wondering if the keto diet is right for you, our registered dietitians explain the eating plan and its pros and cons. What is the keto diet? The ketogenic or keto diet includes eating high-fat, low-carbohydrate foods to reach ketosis. In ketosis, the body uses fat for fuel, instead of sugar (glucose). This produces acids known as ketones. What are the basic rules of the keto diet? There are several different styles of keto diets. Each calls for eating different amounts of fat, protein and carbs in a day. Classic Keto Diet This is the original and best-studied of the different diet strategies. It's recommended that you work with a registered dietitian as it's the strictest, most difficult diet to follow. This diet is very in high fat, very low in carbs and relatively low in protein. All foods must be weighed on a scale to achieve a relative weight distribution of 4 grams of fat for every 1 gram of combined protein and carb eaten throughout the day. In other words, you will likely need to avoid all carb-rich foods (like pasta, potatoes, fruit, etc.). Also, for every serving of protein-rich food, you need four servings of a fatty food (heavy cream, avocado, olive oil, full fat cheese, etc.). A typical meal might look like half an avocado with a small piece of salmon cooked in a lot of olive oil with a side salad. It will all be topped with a cream sauce made with heavy cream and cheese. You would need to eat all of the oils and creams. Modified Keto Diet This is a modified form of the classic keto diet, allowing slightly more protein and carbs. This diet generally still requires a scale for weighing food. It's defined by a ratio closer to 3 grams of fat for every 1 gram of combined protein and carb (generally with a focus on more protein than carbs). In other words, you'll still avoid most carb-rich foods except for maybe one or two servings per day. For every serving of a protein- or carb-rich food, you'll need three servings of very high-fat food. Modified Atkins Diet This diet is the more user-friendly keto diet and likely the most popular. It's a high-fat, low-carb diet where you can choose to eat as much protein as you like. What foods can you eat on the keto diet? Here's a list of foods you should eat on the keto diet: Dairy and unsweetened dairy alternatives such as full fat cheese, plain full fat yogurt, cottage cheese and unsweetened plant-based milk. Low-carb vegetables like leafy greens, peppers and broccoli. Other plant-based foods including nuts, seeds and avocados. Oils and fats like olive and coconut oils, as well as whole olives and avocados. Moderate amounts of high protein foods such as seafood, meat, poultry and eggs. Sometimes small amounts of berries can be eaten. These diets generally avoid all starchy foods (pasta, rice, potatoes, bread, corn, peas, beans and legumes). Desserts and any other high-carb foods are also not part of the diet. Fruits can sometimes be included if they are high in fiber, such as raspberries. What are the pros and cons of the keto diet? Here are some benefits of the keto diet for your health: Helps with weight loss: Research suggests that the keto diet helps people cut calories because the diet doesn't include many foods that people are likely to overeat. However, it can be difficult to transition off of the keto diet without regaining weight. That's because people typically start eating more high-carb foods. Some data supports going from the keto diet to the Mediterranean diet or MIND diet may help maintain healthier weight once weight loss goals are achieved. Benefits people with type 2 diabetes: Some people with diabetes find that reducing their carb intake can help with blood sugar and medication management. Epilepsy management: Data shows that the keto diet can have positive effects for people with epilepsy in helping with seizure management. Some cons of the keto diet include: Can lead to nutrient deficiencies: The keto diet is very low in carbs which means you can't eat many fruits and vegetables. That can lead to low levels of vitamins A, E and B6, as well as folate, calcium, magnesium, and potassium. Possible negative effects on heart health: There are mixed results in studies that examine the specific effects of the keto diet on heart health. This is likely because much of it depends on individual food choices. Generally, foods high in saturated fat and a high-fat diet have negative consequences on heart health factors such as your lipid panel. Constipation and other gastrointestinal (GI) troubles: Constipation is likely the greatest long-term complication of the keto diet because of a low fiber intake. Fiber is a form of carb that is digested not by our human cells, but by the microbes in our gut. Any high-fiber foods (like beans, whole grains, and certain fruits) are also high in carbs, which is not typically part of the keto diet. An ongoing diet that is low in fiber (including the typical Western diet) may leave you at higher risk for certain cancers like colon cancer. What are some side effects of the keto diet? Before starting the keto diet, you should get a blood test to look at your lipids. You should then take another blood test after three to six months on the keto diet. Some people will actually find that their high-density lipoproteins (HDL) – the good cholesterol – go up and their triglycerides go down. But some people won't see a positive shift in those markers and may find their low-density lipoproteins (LDL) – the bad cholesterol – go up to unhealthy levels. There can be issues with bone health when people switch to low-carb keto diets. Researchers see that markers for bone breakdown are higher and markers for bone building are lower. This is a common side effect in kids who are fed the keto diet to treat epilepsy. Also, when people start the keto diet, they go through the keto flu. Keto flu symptoms appear a few days after starting the diet. This eventually goes away once your body adjusts to your eating changes. People on the keto diet also tend to lose more salt/sodium than those on a higher carb diet and may need more salt than the average person. Most people following a strict keto diet should take a multivitamin with minerals to ensure they're meeting those vitamin and mineral needs. The diet itself generally is not enough. Does the keto diet improve brain functions, such as learning and memory? The keto diet can show positive results for helping the brain. It's been used for years to treat epilepsy. Researchers are looking to see if the keto diet can help treat age-related brain decline and potentially Alzheimer's disease. There are mixed results on if the keto diet can help those with neuromuscular diseases such as Parkinson's disease. There are many different theories as to why the keto diet may be beneficial for brain health. These include the direct benefit from the presence of ketone bodies, especially one called beta hydroxybutyrate. Other theories suggest a possible benefit in changing the microbiome for certain conditions, like epilepsy. Is the keto diet good for athletes? Our bodies use multiple energy systems to exercise. The type of energy system used depends on the exercise intensity and duration. At lower intensities, our bodies use a mix of both burning fat and carbs for fuel. When we transition to high intensities (such as a scheduled workout), our bodies start to rely primarily on carbs. Most studies involving endurance athletes showed the keto diet hurts their performance. The keto diet has also been shown to increase losses of lean tissue when compared to a higher-carb diet. The bottom line is that the keto diet impairs athletic performance and are not recommended for athletes because they restrict carbs. What other eating plans might be right for me? If the keto diet doesn't sound like the right fit for you, you may want to consider one of these other eating plans. Intermittent fasting This weight-loss trend dates back to ancient times where fasting helps heal the body. The idea behind intermittent fasting is that by restricting food, our bodies will more quickly tap into fat stores for energy. It's not necessarily an eating plan that outlines what types of foods to eat but instead focuses on when to eat. There are a couple of different ways to implement intermittent fasting. Mediterranean diet This way of eating focuses on whole, plant-based foods and healthy fat. The Mediterranean diet's methods are scientifically proven to reduce the risk of chronic conditions like heart disease. MIND Diet Short for "Mediterranean-DASH Intervention for Neurodegenerative Delay," this diet is similar to the Mediterranean diet, but with stricter requirements for what you can eat. It emphasizes eating more vegetables, fruits (specifically berries), high-fiber foods, beans, nuts, seafood, poultry and olive oil. The MIND diet can possibly help reduce the cognitive decline associated with aging. It also may help slow disease progression for conditions like Parkinson's disease.
考点1: “keto diet”应该译为“生酮饮食”。 考点2: ”cottage cheese”应译为“茅屋奶酪”或“考特吉奶酪”或“乡村奶酪” 考点3: “MIND diet”推荐翻译为“麦得”或“Mind饮食” 考点4:“Mediterranean-DASH Intervention for Neurodegenerative Delay”推荐译为“地中海-DASH干预延迟神经变性饮食”
66253
垂类场景
食品健康
68
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: AI视频生成技术正成为连接古今、沟通中外的崭新纽带,让尘封的历史记忆焕发出新的活力,也让文化的交流更加生动可及。 如果陈子昂活在今天,或许不需要怆然涕下了——近日,“AI穿越直播”火了。 视频中,AI记者“穿越”至古代场景:或在烽火连天的赤壁战场奔走解说,或“现身”唐朝玄武门之变现场,或“参与”秦朝长城修筑……这样的作品如雨后春笋般涌现,凭借“第一人称Vlog视角+历史名场面”的独特叙事,营造出强烈的“在场感”,让观众在虚实交融中沉浸式感受历史的鲜活脉搏。 那些耳熟能详的历史事件,以前所未有的方式“活”了起来,这要得益于AI视频生成技术的突破性发展。技术手段有效弥合了时空隔阂带来的疏离感,今人才能通过AI视频“亲历”那些惊心动魄的名场面。这种深度沉浸的体验,精准回应了人们对历史的渴求——历史不再是书本上静止的文字和图片,而变得可感、可知,仿佛触手可及,成为我们理解过去的一把钥匙。 这种鲜活的科技呈现方式,为文化传播架起了一座新桥梁。对青少年学生而言,AI视频以其生动的画面和沉浸的视角,将课本上的知识变得引人入胜。比如,动态重现的“玄武门之变”,远比文字描述更能点燃他们探究历史的兴趣。对海外观众来说,配备多语种的AI历史短片,成为他们了解中国历史与文化的直观窗口,让中国故事更易抵达全球受众。可以说,AI视频生成技术正成为连接古今、沟通中外的崭新纽带,让尘封的历史记忆焕发出新的活力,也让文化的交流更加生动可及。 更重要的是,这种AI视频赋予了普通人前所未有的创作自由。曾几何时,制作一部穿越历史的影视作品,需要耗费大量人力物力。如今,数字技术的普及大大降低了创作门槛,众多网友得以亲身参与。他们不再是单向的受众,而是拿起AI工具,成为历史文化内容的创作者和传播者,能动地参与到讲述历史、传承文化的潮流中来。这种技术对创造力的极大解放,让更多元的视角、更多样的故事得以更新奇地呈现,为文化的繁荣发展注入了源源不断的新活力。 技术是把双刃剑,热闹之下也有隐忧。如果说,网上冒出一些让人哭笑不得的内容,比如“秦始皇直播卖瓷砖”“屈原代言粽子”,将历史娱乐化,还能让人一眼看破、一笑置之的话,有些AI作品把不同历史时期的东西凑在一起,却又“一本正经”地讲述,则有可能误导儿童青少年,容易将AI生成的“历史八卦”当成真实历史,失去对历史的敬畏感。 技术在进步,人也需要进步。技术迭代向前的每一步,都呼唤着社会认知与法规体系的同步演进。提高AI素养,教育系统已开始行动。例如,北京中小学从今年秋天起,将把AI素养纳入课程体系,培育批判思维与责任伦理并重的技术观。另一方面,监管也在跟进,《生成式人工智能服务管理暂行办法》明确要求标识合成内容,广电等部门也曾对歪曲历史、误导公众的“AI魔改”视频开展专项整治。 当然,技术创新用对了地方,就是好事。君不见,浙江温州用AI“复活”了宋元时期古港千帆竞发的盛景,让沉寂的码头重新响起商贾云集的喧嚣;苏州博物馆借助AI技术“复活”唐伯虎并推出“回答我”系列视频,让网友直呼过瘾;各地博物馆应用AI、VR(虚拟现实)、裸眼3D等技术,打破时空壁垒,让沉睡千年的文物“开口讲故事”……这说明,只要我们心里装着对历史的尊重,用好技术这把“钥匙”,就能打开创新传承的新大门。
AI视频生成技术正成为连接古今、沟通中外的崭新纽带,让尘封的历史记忆焕发出新的活力,也让文化的交流更加生动可及。 如果陈子昂活在今天,或许不需要怆然涕下了——近日,“AI穿越直播”火了。 视频中,AI记者“穿越”至古代场景:或在烽火连天的赤壁战场奔走解说,或“现身”唐朝玄武门之变现场,或“参与”秦朝长城修筑……这样的作品如雨后春笋般涌现,凭借“第一人称Vlog视角+历史名场面”的独特叙事,营造出强烈的“在场感”,让观众在虚实交融中沉浸式感受历史的鲜活脉搏。 那些耳熟能详的历史事件,以前所未有的方式“活”了起来,这要得益于AI视频生成技术的突破性发展。技术手段有效弥合了时空隔阂带来的疏离感,今人才能通过AI视频“亲历”那些惊心动魄的名场面。这种深度沉浸的体验,精准回应了人们对历史的渴求——历史不再是书本上静止的文字和图片,而变得可感、可知,仿佛触手可及,成为我们理解过去的一把钥匙。 这种鲜活的科技呈现方式,为文化传播架起了一座新桥梁。对青少年学生而言,AI视频以其生动的画面和沉浸的视角,将课本上的知识变得引人入胜。比如,动态重现的“玄武门之变”,远比文字描述更能点燃他们探究历史的兴趣。对海外观众来说,配备多语种的AI历史短片,成为他们了解中国历史与文化的直观窗口,让中国故事更易抵达全球受众。可以说,AI视频生成技术正成为连接古今、沟通中外的崭新纽带,让尘封的历史记忆焕发出新的活力,也让文化的交流更加生动可及。 更重要的是,这种AI视频赋予了普通人前所未有的创作自由。曾几何时,制作一部穿越历史的影视作品,需要耗费大量人力物力。如今,数字技术的普及大大降低了创作门槛,众多网友得以亲身参与。他们不再是单向的受众,而是拿起AI工具,成为历史文化内容的创作者和传播者,能动地参与到讲述历史、传承文化的潮流中来。这种技术对创造力的极大解放,让更多元的视角、更多样的故事得以更新奇地呈现,为文化的繁荣发展注入了源源不断的新活力。 技术是把双刃剑,热闹之下也有隐忧。如果说,网上冒出一些让人哭笑不得的内容,比如“秦始皇直播卖瓷砖”“屈原代言粽子”,将历史娱乐化,还能让人一眼看破、一笑置之的话,有些AI作品把不同历史时期的东西凑在一起,却又“一本正经”地讲述,则有可能误导儿童青少年,容易将AI生成的“历史八卦”当成真实历史,失去对历史的敬畏感。 技术在进步,人也需要进步。技术迭代向前的每一步,都呼唤着社会认知与法规体系的同步演进。提高AI素养,教育系统已开始行动。例如,北京中小学从今年秋天起,将把AI素养纳入课程体系,培育批判思维与责任伦理并重的技术观。另一方面,监管也在跟进,《生成式人工智能服务管理暂行办法》明确要求标识合成内容,广电等部门也曾对歪曲历史、误导公众的“AI魔改”视频开展专项整治。 当然,技术创新用对了地方,就是好事。君不见,浙江温州用AI“复活”了宋元时期古港千帆竞发的盛景,让沉寂的码头重新响起商贾云集的喧嚣;苏州博物馆借助AI技术“复活”唐伯虎并推出“回答我”系列视频,让网友直呼过瘾;各地博物馆应用AI、VR(虚拟现实)、裸眼3D等技术,打破时空壁垒,让沉睡千年的文物“开口讲故事”……这说明,只要我们心里装着对历史的尊重,用好技术这把“钥匙”,就能打开创新传承的新大门。
考点1:【多语种AI历史短片】 应译为 【multilingual AI historical short videos】 考点2:【AI魔改】 应译为【 AI-modified content】 考点3:【裸眼3D】 应译为 【glasses-free 3D】 考点4:【开口讲故事】 应译为 【make relics “speak” their stories】
67078
新闻资讯
观点评论
91
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Genome-wide analysis identifies molecular systems and 149 genetic loci associated with income Abstract Socioeconomic position (SEP) is a multi-dimensional construct reflecting (and influencing) multiple socio-cultural, physical, and environmental factors. In a sample of 286,301 participants from UK Biobank, we identify 30 (29 previously unreported) independent-loci associated with income. Using a method to meta-analyze data from genetically-correlated traits, we identify an additional 120 income-associated loci. These loci show clear evidence of functionality, with transcriptional differences identified across multiple cortical tissues, and links to GABAergic and serotonergic neurotransmission. By combining our genome wide association study on income with data from eQTL studies and chromatin interactions, 24 genes are prioritized for follow up, 18 of which were previously associated with intelligence. We identify intelligence as one of the likely causal, partly-heritable phenotypes that might bridge the gap between molecular genetic inheritance and phenotypicconsequence in terms of income differences. These results indicate that, in modern era Great Britain, genetic effects contribute towards some of the observed socioeconomic inequalities. Introduction People with advantaged socioeconomic backgrounds, on average, live longer, and have better mental and physical health than those from more deprived environments. An understanding of the causes underlying the association between socioeconomic position (SEP) and health is likely to be helpful to minimize social disparities in health and well-being. The link between SEP and health is typically thought to be due to environmental factors, including, but not limited to access to resources, exposure to harmful or stressful environments, adverse health behaviours, such as smoking, poor diet and excessive alcohol consumption, and a lack of physical exercise5. However, genetic factors (most likely via mediated pleiotropy, Fig. 1) have been discussed as a partial explanation for the SEP–health gradient; for example, genetic predispositions towards certain diseases and/or genetic influences on what foods people like, could lead to poor diet, which in turn could lead to both lower SEP and poorer health6. It has recently been demonstrated that genome-wide association studies (GWASs) can capture shared genetic associations between measures of health and SEP7. Potential pleiotropic effects are highlighted in the observed genetic correlations between SEP variables, such as years of education completed, household income and social deprivation, and physical and mental health traits including longevity. Loci associated with two SEP phenotypes, education and household income, have been identified via GWASs7,9,10,11, but—consistent with other complex traits, such as height—these loci collectively account for only a small fraction of the total heritability of the traits in question. For household income, an analysis of a sample of 96,900 individuals from Great Britain found that additive genetic effects tagged by common single-nucleotide polymorphisms (SNPs) accounted for ~11% (SE = 0.7%) of differences in household income7. Two loci attained genome-wide significance in that study, but they collectively accounted for <0.005% of the total SNP heritability. Here, we use the UK Biobank data set12 to examine genetic associations with household income (n = 286,301) in a contemporary British sample. We identify 30 independent genome-wide significant loci, 29 of which are unreported in previous work. Using a method that leverages power from genetically correlated traits, multi-trait analysis of GWAS (MTAG), an additional 120 loci are found to be associated with income. We identify neurogenesis and the components of the synapse as being associated with income. Furthermore, we link transcription differences across multiple cortical tissue types, as well as both GABAergic and serotonergic neurotransmission, to income differences. We also show that the genes linked to differences in income are predominantly those that have been previously linked with intelligence8, and that intelligence is one of the likely causal factors leading to differences in income. We compare the genetic correlations derived using income with those derived using another measure of SEP, educational attainment, to show that the genetic variants associated with income are related to better mental health than those related to education. Finally, we predict 2.5% of income differences using genetic data alone in an independent sample. Results Graphical representation of statistical analysis A flow chart summarizing all statistical analyses conducted is displayed in Fig. 2. SNP-based analysis of income For household income, 3712 SNPs attained genome-wide significance (P < 5 × 10−8), across 30 independent loci (Fig. 3a and Supplementary Data 1), which contained 68 independent significant SNPs and 31 lead SNPs. A total of 29 of these 30 loci were not identified in the previous UK Biobank analysis of income (Supplementary Data 2). The 30 loci predominantly contained SNPs found within intronic regions (47%) as well as non-coding RNA introns (29%). A total of 17% of the SNPs within the independent loci were found in intergenic regions, and only 1.2% were found in exons (Fig. 3b). Many of the loci contained SNPs showing evidence of influencing gene regulation with 33% having a Regulome DB score of <2 (Fig. 3c) and 86% having evidence of being in an open-chromatin state (indicated by a minimum chromatin state of <8, in Fig. 3d). Additionally, these loci were linked to intelligence (11 loci), mental health (schizophrenia, 1 locus; bipolar disorder, 2 loci; neuroticism, 4 loci) and neurological variables (corticobasal degeneration, 1 locus; subcortical brain volumes, 1 locus; and Parkinson’s disease, 1 locus) (Supplementary Data 3). Linkage disequilibrium score (LDSC) regression showed that the mean χ2 statistic was 1.45 and the intercept of the LDSC regression was 1.04. These statistics indicate that around 92% of the inflation in the GWAS test statistics was due to a polygenic signal rather than residual stratification or confounding. The LDSC regression estimate of the heritability of household income was 7.39% (SE = 0.33%). Gene prioritization Three methods of mapping allelic variation to genes were used to better understand the functional consequences of the 30 independent loci linked to income (positional mapping, expression quantitative trait loci (eQTL) analysis and chromatin mapping). Using positional mapping, SNPs from the GWAS were aligned to 117 genes. eQTL mapping was used to match cis-eQTL SNPs to 186 genes, and chromatin interaction mapping linked the SNPs to 277 genes (Fig. 3e and Supplementary Data 4 and Supplementary Fig. 1). These mapping strategies identified a total of 400 unique genes, of which 133 (Fig. 3e cells 14 + 23 + 26 + 3 + 24 + 11 + 2 + 30) were implicated by at least two mapping strategies and 47 (Fig. 3e cells 23 + 24) were implicated by all three. Of the 133 implicated by two mapping strategies, two showed evidence of a chromatin interactions with two independent genomic risk loci (Supplementary Data 5). Both HOXB2 and HOXB7 showed interactions with loci 24 and loci 25. HOXB2 showed interactions in mesendoderm (an embryonic tissue layer) tissue and IMR90 (foetal lung fibroblasts)tissue, whereas HOXB7 showed associations in the tissues of hESC (human embryonic stem cell), mesenchymal (multipotent stromal cells which differentiate into a variety of different cell types) stem cell, IMR90, left ventricle, GM12878, and trophoblast-like cells.
Genome-wide analysis identifies molecular systems and 149 genetic loci associated with income Abstract Socioeconomic position (SEP) is a multi-dimensional construct reflecting (and influencing) multiple socio-cultural, physical, and environmental factors. In a sample of 286,301 participants from UK Biobank, we identify 30 (29 previously unreported) independent-loci associated with income. Using a method to meta-analyze data from genetically-correlated traits, we identify an additional 120 income-associated loci. These loci show clear evidence of functionality, with transcriptional differences identified across multiple cortical tissues, and links to GABAergic and serotonergic neurotransmission. By combining our genome wide association study on income with data from eQTL studies and chromatin interactions, 24 genes are prioritized for follow up, 18 of which were previously associated with intelligence. We identify intelligence as one of the likely causal, partly-heritable phenotypes that might bridge the gap between molecular genetic inheritance and phenotypicconsequence in terms of income differences. These results indicate that, in modern era Great Britain, genetic effects contribute towards some of the observed socioeconomic inequalities. Introduction People with advantaged socioeconomic backgrounds, on average, live longer, and have better mental and physical health than those from more deprived environments. An understanding of the causes underlying the association between socioeconomic position (SEP) and health is likely to be helpful to minimize social disparities in health and well-being. The link between SEP and health is typically thought to be due to environmental factors, including, but not limited to access to resources, exposure to harmful or stressful environments, adverse health behaviours, such as smoking, poor diet and excessive alcohol consumption, and a lack of physical exercise5. However, genetic factors (most likely via mediated pleiotropy, Fig. 1) have been discussed as a partial explanation for the SEP–health gradient; for example, genetic predispositions towards certain diseases and/or genetic influences on what foods people like, could lead to poor diet, which in turn could lead to both lower SEP and poorer health6. It has recently been demonstrated that genome-wide association studies (GWASs) can capture shared genetic associations between measures of health and SEP7. Potential pleiotropic effects are highlighted in the observed genetic correlations between SEP variables, such as years of education completed, household income and social deprivation, and physical and mental health traits including longevity. Loci associated with two SEP phenotypes, education and household income, have been identified via GWASs7,9,10,11, but—consistent with other complex traits, such as height—these loci collectively account for only a small fraction of the total heritability of the traits in question. For household income, an analysis of a sample of 96,900 individuals from Great Britain found that additive genetic effects tagged by common single-nucleotide polymorphisms (SNPs) accounted for ~11% (SE = 0.7%) of differences in household income7. Two loci attained genome-wide significance in that study, but they collectively accounted for <0.005% of the total SNP heritability. Here, we use the UK Biobank data set12 to examine genetic associations with household income (n = 286,301) in a contemporary British sample. We identify 30 independent genome-wide significant loci, 29 of which are unreported in previous work. Using a method that leverages power from genetically correlated traits, multi-trait analysis of GWAS (MTAG), an additional 120 loci are found to be associated with income. We identify neurogenesis and the components of the synapse as being associated with income. Furthermore, we link transcription differences across multiple cortical tissue types, as well as both GABAergic and serotonergic neurotransmission, to income differences. We also show that the genes linked to differences in income are predominantly those that have been previously linked with intelligence8, and that intelligence is one of the likely causal factors leading to differences in income. We compare the genetic correlations derived using income with those derived using another measure of SEP, educational attainment, to show that the genetic variants associated with income are related to better mental health than those related to education. Finally, we predict 2.5% of income differences using genetic data alone in an independent sample. Results Graphical representation of statistical analysis A flow chart summarizing all statistical analyses conducted is displayed in Fig. 2. SNP-based analysis of income For household income, 3712 SNPs attained genome-wide significance (P < 5 × 10−8), across 30 independent loci (Fig. 3a and Supplementary Data 1), which contained 68 independent significant SNPs and 31 lead SNPs. A total of 29 of these 30 loci were not identified in the previous UK Biobank analysis of income (Supplementary Data 2). The 30 loci predominantly contained SNPs found within intronic regions (47%) as well as non-coding RNA introns (29%). A total of 17% of the SNPs within the independent loci were found in intergenic regions, and only 1.2% were found in exons (Fig. 3b). Many of the loci contained SNPs showing evidence of influencing gene regulation with 33% having a Regulome DB score of <2 (Fig. 3c) and 86% having evidence of being in an open-chromatin state (indicated by a minimum chromatin state of <8, in Fig. 3d). Additionally, these loci were linked to intelligence (11 loci), mental health (schizophrenia, 1 locus; bipolar disorder, 2 loci; neuroticism, 4 loci) and neurological variables (corticobasal degeneration, 1 locus; subcortical brain volumes, 1 locus; and Parkinson’s disease, 1 locus) (Supplementary Data 3). Linkage disequilibrium score (LDSC) regression showed that the mean χ2 statistic was 1.45 and the intercept of the LDSC regression was 1.04. These statistics indicate that around 92% of the inflation in the GWAS test statistics was due to a polygenic signal rather than residual stratification or confounding. The LDSC regression estimate of the heritability of household income was 7.39% (SE = 0.33%). Gene prioritization Three methods of mapping allelic variation to genes were used to better understand the functional consequences of the 30 independent loci linked to income (positional mapping, expression quantitative trait loci (eQTL) analysis and chromatin mapping). Using positional mapping, SNPs from the GWAS were aligned to 117 genes. eQTL mapping was used to match cis-eQTL SNPs to 186 genes, and chromatin interaction mapping linked the SNPs to 277 genes (Fig. 3e and Supplementary Data 4 and Supplementary Fig. 1). These mapping strategies identified a total of 400 unique genes, of which 133 (Fig. 3e cells 14 + 23 + 26 + 3 + 24 + 11 + 2 + 30) were implicated by at least two mapping strategies and 47 (Fig. 3e cells 23 + 24) were implicated by all three. Of the 133 implicated by two mapping strategies, two showed evidence of a chromatin interactions with two independent genomic risk loci (Supplementary Data 5). Both HOXB2 and HOXB7 showed interactions with loci 24 and loci 25. HOXB2 showed interactions in mesendoderm (an embryonic tissue layer) tissue and IMR90 (foetal lung fibroblasts)tissue, whereas HOXB7 showed associations in the tissues of hESC (human embryonic stem cell), mesenchymal (multipotent stromal cells which differentiate into a variety of different cell types) stem cell, IMR90, left ventricle, GM12878, and trophoblast-like cells.
考点1: "genetic loci"应该译为“基因座”。 考点2: "meta-analyze"应译为“统合分析”。 考点3: "GABAergic '应译为“迦瑪胺基丁酸”。 考点4:" eQTL" 应该译为“表达数量性状位点”。 考点5:" pleiotropy'应译为“基因多效性”。 考点6: "synapse'应译为“突触”。 考点7:"exons'应译为“外显子”。 考点8:"RNA'应译为“核糖核酸”。 考点9:"HOXB2"应译为“同源核基因B2”。 考点10:"Linkage disequilibrium score"应译为“连锁不平衡分数”。 考点11:'foetal lung fibroblasts"应译为“胎儿肺成纤维细胞 ”,其中foetal应该是拼写错误。
677ce
学术论文
自然科学
67
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 从"北大废物"到暴瘦女神:李雪琴用30斤体重,找回了“自己” 谁能想到,那个曾经胖乎乎的李雪琴,如今竟成了众人口中的 " 美貌女神 "? 2025 年,当她再次出现在《你好星期六》的舞台上时,观众们几乎认不出眼前这个气质出众的女子。 腰线纤细,皮肤紧致,网友都震惊了:" 这还是我们认识的那个李雪琴吗?" 三十斤的体重,她说减就减了。 她曾在节目中坦言,暗恋的男生因为她胖,随口说出 " 瘦到一百斤才能抱你 " 这样的话。 那时的她卑微到尘埃里,自尊心支离破碎。 而现在,她甚至被传有了同样高学历背景的男友。据说此人是《宇宙探索编辑部》的编剧,在马路上牵着她的手,非常甜蜜。 这么多年过去了,她终于不再是那个被人嘲笑 " 没人抱 " 的女孩,开始闪闪发光。 网红时代,意外爆红 2019 年夏天,李雪琴走红了。 她穿着一件普通 T 恤,杵在清华校门前,立好了拍摄的设备。随后便向某明星喊话:" 你看这清华大学的校门,多白!" 就是这样一个简单到近乎粗糙的视频,突然在网络上火了。 人们议论着这个女孩,说她跟其他网红一样装疯卖傻。 质疑声此起彼伏:" 顶尖学校毕业的精英竟然拍这种低趣味视频,简直浪费了国家的栽培。" 面对质疑,她淡然回应:" 北大就不能养废物吗?" 这种自嘲式的反击,反而让她获得了更多关注。 用最朴实的方式展现了普通人的真实状态,这种反差感恰好击中了人们的审美疲劳。 没有精致滤镜、华丽包装,她给出了最接地气的表达、最真实的自己。 这种真实,成了她最大的武器。 脱口秀舞台,天才绽放 2020 年,她以 " 脱口秀天才少女 " 的身份再次火出圈。 舞台上的她,把生活中的平淡日常改写成了一个个让人哈哈大笑的段子。 吐槽老板半夜三点打电话,说她这大半夜的怎么还睡觉;吐槽妈妈催婚,结果妈妈自己先再婚了。 " 宇宙的尽头是铁岭 " 这句话,从她口中说出后迅速成为网络金句。 很多人觉得她应该是个乐天派,殊不知喜剧的内核是悲剧。 因为成长背景和学业压力,她曾经患过抑郁症,幽默不过是解构痛苦的一种方式。 铁岭女孩,敏感童年 1995 年,她出生在东北铁岭。 小时候她就是个“圆球”,长得也不上相。上小学时,很多同学因为她的身材嘲笑孤立她,敏感的她因此变得自卑内向。 为了弥补身材和长相的不足,她在学习上特别上进。每次考试都能拿第一,因此获得了 " 李第一 " 的绰号。 14 岁那年,做生意失败的父亲选择离开这个家,原本幸福的三口之家坍塌了。母亲情绪变得极度不稳定,芝麻小事也有可能突然乱发脾气。 " 我妈是一个小女生,她是我带着长大的,这个家是我撑起来的。" 她练就了“高情商”,不自觉地就会安抚别人的情绪。 让妈妈开心,成了她做事的唯一准则。 有一次考试她只拿了第二名,邻居无意说了一句 " 可能是受父母离婚影响 ",母亲当场崩溃痛哭。从此之后,她每次考试都是第一,最终考入北大。 可是,光顾着治愈母亲情绪的她,没有意识到自己的情绪也 " 生病 " 了,内心的快乐却越来越少。 她曾经到纽约大学继续求学之路。但是陌生的环境和巨大的压力让她的抑郁症愈发严重。 她经常整日瘫在床上,望着天花板发呆流泪,直到精疲力竭地睡去。 走在国外的大街上,也会止不住嚎啕大哭。最严重的时候,她甚至用小刀在手腕上划下伤痕。 最终,学业无法继续的她选择了退学。有时候,退一步不是懦弱,而是为了更好地找准方向。 学新闻的她,对网络平台很感兴趣,便开始尝试短视频。谁知一发不可收拾,走上了网红之路,深受网友欢迎。 而这些努力,也让她渐渐走上了治愈自己的道路。自此拍综艺、讲脱口秀、演戏,一发不可收拾。 从铁岭的胖女孩到到网络的搞笑红人,她用三十年的时间完成了一场漫长的自我和解。 人生从来不是一条直线,跌跌撞撞才是常态。 重要的是,无论走过多少弯路,都要记得最初的那个自己,那个敢于做梦、敢于发光的自己。 如今的她,终于活成了自己想要的样子——有才华的光芒,更有内心的平静与自信。
从"北大废物"到暴瘦女神:李雪琴用30斤体重,找回了“自己” 谁能想到,那个曾经胖乎乎的李雪琴,如今竟成了众人口中的 " 美貌女神 "? 2025 年,当她再次出现在《你好星期六》的舞台上时,观众们几乎认不出眼前这个气质出众的女子。 腰线纤细,皮肤紧致,网友都震惊了:" 这还是我们认识的那个李雪琴吗?" 三十斤的体重,她说减就减了。 她曾在节目中坦言,暗恋的男生因为她胖,随口说出 " 瘦到一百斤才能抱你 " 这样的话。 那时的她卑微到尘埃里,自尊心支离破碎。 而现在,她甚至被传有了同样高学历背景的男友。据说此人是《宇宙探索编辑部》的编剧,在马路上牵着她的手,非常甜蜜。 这么多年过去了,她终于不再是那个被人嘲笑 " 没人抱 " 的女孩,开始闪闪发光。 网红时代,意外爆红 2019 年夏天,李雪琴走红了。 她穿着一件普通 T 恤,杵在清华校门前,立好了拍摄的设备。随后便向某明星喊话:" 你看这清华大学的校门,多白!" 就是这样一个简单到近乎粗糙的视频,突然在网络上火了。 人们议论着这个女孩,说她跟其他网红一样装疯卖傻。 质疑声此起彼伏:" 顶尖学校毕业的精英竟然拍这种低趣味视频,简直浪费了国家的栽培。" 面对质疑,她淡然回应:" 北大就不能养废物吗?" 这种自嘲式的反击,反而让她获得了更多关注。 用最朴实的方式展现了普通人的真实状态,这种反差感恰好击中了人们的审美疲劳。 没有精致滤镜、华丽包装,她给出了最接地气的表达、最真实的自己。 这种真实,成了她最大的武器。 脱口秀舞台,天才绽放 2020 年,她以 " 脱口秀天才少女 " 的身份再次火出圈。 舞台上的她,把生活中的平淡日常改写成了一个个让人哈哈大笑的段子。 吐槽老板半夜三点打电话,说她这大半夜的怎么还睡觉;吐槽妈妈催婚,结果妈妈自己先再婚了。 " 宇宙的尽头是铁岭 " 这句话,从她口中说出后迅速成为网络金句。 很多人觉得她应该是个乐天派,殊不知喜剧的内核是悲剧。 因为成长背景和学业压力,她曾经患过抑郁症,幽默不过是解构痛苦的一种方式。 铁岭女孩,敏感童年 1995 年,她出生在东北铁岭。 小时候她就是个“圆球”,长得也不上相。上小学时,很多同学因为她的身材嘲笑孤立她,敏感的她因此变得自卑内向。 为了弥补身材和长相的不足,她在学习上特别上进。每次考试都能拿第一,因此获得了 " 李第一 " 的绰号。 14 岁那年,做生意失败的父亲选择离开这个家,原本幸福的三口之家坍塌了。母亲情绪变得极度不稳定,芝麻小事也有可能突然乱发脾气。 " 我妈是一个小女生,她是我带着长大的,这个家是我撑起来的。" 她练就了“高情商”,不自觉地就会安抚别人的情绪。 让妈妈开心,成了她做事的唯一准则。 有一次考试她只拿了第二名,邻居无意说了一句 " 可能是受父母离婚影响 ",母亲当场崩溃痛哭。从此之后,她每次考试都是第一,最终考入北大。 可是,光顾着治愈母亲情绪的她,没有意识到自己的情绪也 " 生病 " 了,内心的快乐却越来越少。 她曾经到纽约大学继续求学之路。但是陌生的环境和巨大的压力让她的抑郁症愈发严重。 她经常整日瘫在床上,望着天花板发呆流泪,直到精疲力竭地睡去。 走在国外的大街上,也会止不住嚎啕大哭。最严重的时候,她甚至用小刀在手腕上划下伤痕。 最终,学业无法继续的她选择了退学。有时候,退一步不是懦弱,而是为了更好地找准方向。 学新闻的她,对网络平台很感兴趣,便开始尝试短视频。谁知一发不可收拾,走上了网红之路,深受网友欢迎。 而这些努力,也让她渐渐走上了治愈自己的道路。自此拍综艺、讲脱口秀、演戏,一发不可收拾。 从铁岭的胖女孩到到网络的搞笑红人,她用三十年的时间完成了一场漫长的自我和解。 人生从来不是一条直线,跌跌撞撞才是常态。 重要的是,无论走过多少弯路,都要记得最初的那个自己,那个敢于做梦、敢于发光的自己。 如今的她,终于活成了自己想要的样子——有才华的光芒,更有内心的平静与自信。
考点1:“废物”应该译为“loser”。 考点2: “你好星期六”应译为“Hello Saturday”或“H!6”。 考点3: “气质出众”应译为“exceptional poise”。 考点4: “宇宙探索编辑部”应译为“Journey to the West”。 考点5: “装疯卖傻”应译为“play dumb and act crazy for attention”。 考点6: “接地气”应译为“unpretencious”。 考点7: “火出圈”应译为“go viral”。 考点8: “金句”应译为“catchphrase”。 考点9:“不上相”应译为“not photogenic”。 考点10:“一发不可收拾”应译为“unstoppable”。 考点11:“弯路中”应译为“setbacks”。
68423
文学艺术
传记
62
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: New method combines imaging and sequencing to study gene function in intact tissue The approach collects multiple types of imaging and sequencing data from the same cells, leading to new insights into mouse liver biology. Imagine that you want to know the plot of a movie, but you only have access to either the visuals or the sound. With visuals alone, you’ll miss all the dialogue. With sound alone, you will miss the action. Understanding our biology can be similar. Measuring one kind of data — such as which genes are being expressed — can be informative, but it only captures one facet of a multifaceted story. For many biological processes and disease mechanisms, the entire “plot” can’t be fully understood without combining data types. However, capturing both the “visuals and sound” of biological data, such as gene expression and cell structure data, from the same cells requires researchers to develop new approaches. They also have to make sure that the data they capture accurately reflects what happens in living organisms, including how cells interact with each other and their environments. Whitehead Institute for Biomedical Research and Harvard University researchers have taken on these challenges and developed Perturb-Multimodal (Perturb-Multi), a powerful new approach that simultaneously measures how genetic changes such as turning off individual genes affect both gene expression and cell structure in intact liver tissue. The method, described in Cell on June 12, aims to accelerate discovery of how genes control organ function and disease. The research team, led by Whitehead Institute Member Jonathan Weissman and then-graduate student in his lab Reuben Saunders, along with Xiaowei Zhuang, the David B. Arnold Professor of Science at Harvard University, and then-postdoc in her lab Will Allen, created a system that can test hundreds of different genetic modifications within a single mouse liver while capturing multiple types of data from the same cells. “Understanding how our organs work requires looking at many different aspects of cell biology at once,” Saunders says. “With Perturb-Multi, we can see how turning off specific genes changes not just what other genes are active, but also how proteins are distributed within cells, how cellular structures are organized, and where cells are located in the tissue. It’s like having multiple specialized microscopes all focused on the same experiment.” “This approach accelerates discovery by both allowing us to test the functions of many different genes at once, and then for each gene, allowing us to measure many different functional outputs or cell properties at once — and we do that in intact tissue from animals,” says Zhuang, who is also a Howard Hughes Medical Institute (HHMI) investigator. A more efficient approach to genetic studies Traditional genetic studies in mice often turn off one gene and then observe what changes in that gene’s absence to learn about what the gene does. The researchers designed their approach to turn off hundreds of different genes across a single liver, while still only turning off one gene per cell — using what is known as a mosaic approach. This allowed them to study the roles of hundreds of individual genes at once in a single individual. The researchers then collected diverse types of data from cells across the same liver to get a full picture of the consequences of turning off the genes. “Each cell serves as its own experiment, and because all the cells are in the same animal, we eliminate the variability that comes from comparing different mice,” Saunders says. “Every cell experiences the same physiological conditions, diet, and environment, making our comparisons much more precise.” “The challenge we faced was that tissues, to perform their functions, rely on thousands of genes, expressed in many different cells, working together. Each gene, in turn, can control many aspects of a cell’s function. Testing these hundreds of genes in mice using current methods would be extremely slow and expensive — near impossible, in practice.” Allen says. Revealing new biology through combined measurements The team applied Perturb-Multi to study genetic controls of liver physiology and function. Their study led to discoveries in three important aspects of liver biology: fat accumulation in liver cells — a precursor to liver disease; stress responses; and hepatocyte zonation (how liver cells specialize, assuming different traits and functions, based on their location within the liver). One striking finding emerged from studying genes that, when disrupted, cause fat accumulation in liver cells. The imaging data revealed that four different genes all led to similar fat droplet accumulation, but the sequencing data showed they did so through three completely different mechanisms. “Without combining imaging and sequencing, we would have missed this complexity entirely,” Saunders says. “The imaging told us which genes affect fat accumulation, while the sequencing revealed whether this was due to increased fat production, cellular stress, or other pathways. This kind of mechanistic insight could be crucial for developing targeted therapies for fatty liver disease.” The researchers also discovered new regulators of liver cell zonation. Unexpectedly, the newly discovered regulators include genes involved in modifying the extracellular matrix — the scaffolding between cells. “We found that cells can change their specialized functions without physically moving to a different zone,” Saunders says. “This suggests that liver cell identity is more flexible than previously thought.” Technical innovation enables new science Developing Perturb-Multi required solving several technical challenges. The team created new methods for preserving the content of interest in cells — RNA and proteins — during tissue processing, for collecting many types of imaging data and single-cell gene expression data from tissue samples that have been fixed with a preservative, and for integrating multiple types of data from the same cells. “Overcoming the inherent complexity of biology in living animals required developing new tools that bridge multiple disciplines — including, in this case, genomics, imaging, and AI,” Allen says. The two components of Perturb-Multi — the imaging and sequencing assays — together, applied to the same tissue, provide insights that are unattainable through either assay alone. “Each component had to work perfectly while not interfering with the others,” says Weissman, who is also a professor of biology at MIT and an HHMI investigator. “The technical development took considerable effort, but the payoff is a system that can reveal biology we simply couldn’t see before.” Expanding to new organs and other contexts The researchers plan to expand Perturb-Multi to other organs, including the brain, and to study how genetic changes affect organ function under different conditions like disease states or dietary changes. “We’re also excited about using the data we generate to train machine learning models,” adds Saunders. “With enough examples of how genetic changes affect cells, we could eventually predict the effects of mutations without having to test them experimentally — a ‘virtual cell’ that could accelerate both research and drug development.” “Perturbation data are critical for training such AI models and the paucity of existing perturbation data represents a major hindrance in such ‘virtual cell’ efforts,” Zhuang says. “We hope Perturb-Multi will fill this gap by accelerating the collection of perturbation data.” The approach is designed to be scalable, with the potential for genome-wide studies that test thousands of genes simultaneously. As sequencing and imaging technologies continue to improve, the researchers anticipate that Perturb-Multi will become even more powerful and accessible to the broader research community. “Our goal is to keep scaling up. We plan to do genome-wide perturbations, study different physiological conditions, and look at different organs,” says Weissman. “That we can now collect so many types of data from so many cells, at speed, is going to be critical for building AI models like virtual cells, and I think it’s going to help us answer previously unsolvable questions about health and disease.”
New method combines imaging and sequencing to study gene function in intact tissue The approach collects multiple types of imaging and sequencing data from the same cells, leading to new insights into mouse liver biology. Imagine that you want to know the plot of a movie, but you only have access to either the visuals or the sound. With visuals alone, you’ll miss all the dialogue. With sound alone, you will miss the action. Understanding our biology can be similar. Measuring one kind of data — such as which genes are being expressed — can be informative, but it only captures one facet of a multifaceted story. For many biological processes and disease mechanisms, the entire “plot” can’t be fully understood without combining data types. However, capturing both the “visuals and sound” of biological data, such as gene expression and cell structure data, from the same cells requires researchers to develop new approaches. They also have to make sure that the data they capture accurately reflects what happens in living organisms, including how cells interact with each other and their environments. Whitehead Institute for Biomedical Research and Harvard University researchers have taken on these challenges and developed Perturb-Multimodal (Perturb-Multi), a powerful new approach that simultaneously measures how genetic changes such as turning off individual genes affect both gene expression and cell structure in intact liver tissue. The method, described in Cell on June 12, aims to accelerate discovery of how genes control organ function and disease. The research team, led by Whitehead Institute Member Jonathan Weissman and then-graduate student in his lab Reuben Saunders, along with Xiaowei Zhuang, the David B. Arnold Professor of Science at Harvard University, and then-postdoc in her lab Will Allen, created a system that can test hundreds of different genetic modifications within a single mouse liver while capturing multiple types of data from the same cells. “Understanding how our organs work requires looking at many different aspects of cell biology at once,” Saunders says. “With Perturb-Multi, we can see how turning off specific genes changes not just what other genes are active, but also how proteins are distributed within cells, how cellular structures are organized, and where cells are located in the tissue. It’s like having multiple specialized microscopes all focused on the same experiment.” “This approach accelerates discovery by both allowing us to test the functions of many different genes at once, and then for each gene, allowing us to measure many different functional outputs or cell properties at once — and we do that in intact tissue from animals,” says Zhuang, who is also a Howard Hughes Medical Institute (HHMI) investigator. A more efficient approach to genetic studies Traditional genetic studies in mice often turn off one gene and then observe what changes in that gene’s absence to learn about what the gene does. The researchers designed their approach to turn off hundreds of different genes across a single liver, while still only turning off one gene per cell — using what is known as a mosaic approach. This allowed them to study the roles of hundreds of individual genes at once in a single individual. The researchers then collected diverse types of data from cells across the same liver to get a full picture of the consequences of turning off the genes. “Each cell serves as its own experiment, and because all the cells are in the same animal, we eliminate the variability that comes from comparing different mice,” Saunders says. “Every cell experiences the same physiological conditions, diet, and environment, making our comparisons much more precise.” “The challenge we faced was that tissues, to perform their functions, rely on thousands of genes, expressed in many different cells, working together. Each gene, in turn, can control many aspects of a cell’s function. Testing these hundreds of genes in mice using current methods would be extremely slow and expensive — near impossible, in practice.” Allen says. Revealing new biology through combined measurements The team applied Perturb-Multi to study genetic controls of liver physiology and function. Their study led to discoveries in three important aspects of liver biology: fat accumulation in liver cells — a precursor to liver disease; stress responses; and hepatocyte zonation (how liver cells specialize, assuming different traits and functions, based on their location within the liver). One striking finding emerged from studying genes that, when disrupted, cause fat accumulation in liver cells. The imaging data revealed that four different genes all led to similar fat droplet accumulation, but the sequencing data showed they did so through three completely different mechanisms. “Without combining imaging and sequencing, we would have missed this complexity entirely,” Saunders says. “The imaging told us which genes affect fat accumulation, while the sequencing revealed whether this was due to increased fat production, cellular stress, or other pathways. This kind of mechanistic insight could be crucial for developing targeted therapies for fatty liver disease.” The researchers also discovered new regulators of liver cell zonation. Unexpectedly, the newly discovered regulators include genes involved in modifying the extracellular matrix — the scaffolding between cells. “We found that cells can change their specialized functions without physically moving to a different zone,” Saunders says. “This suggests that liver cell identity is more flexible than previously thought.” Technical innovation enables new science Developing Perturb-Multi required solving several technical challenges. The team created new methods for preserving the content of interest in cells — RNA and proteins — during tissue processing, for collecting many types of imaging data and single-cell gene expression data from tissue samples that have been fixed with a preservative, and for integrating multiple types of data from the same cells. “Overcoming the inherent complexity of biology in living animals required developing new tools that bridge multiple disciplines — including, in this case, genomics, imaging, and AI,” Allen says. The two components of Perturb-Multi — the imaging and sequencing assays — together, applied to the same tissue, provide insights that are unattainable through either assay alone. “Each component had to work perfectly while not interfering with the others,” says Weissman, who is also a professor of biology at MIT and an HHMI investigator. “The technical development took considerable effort, but the payoff is a system that can reveal biology we simply couldn’t see before.” Expanding to new organs and other contexts The researchers plan to expand Perturb-Multi to other organs, including the brain, and to study how genetic changes affect organ function under different conditions like disease states or dietary changes. “We’re also excited about using the data we generate to train machine learning models,” adds Saunders. “With enough examples of how genetic changes affect cells, we could eventually predict the effects of mutations without having to test them experimentally — a ‘virtual cell’ that could accelerate both research and drug development.” “Perturbation data are critical for training such AI models and the paucity of existing perturbation data represents a major hindrance in such ‘virtual cell’ efforts,” Zhuang says. “We hope Perturb-Multi will fill this gap by accelerating the collection of perturbation data.” The approach is designed to be scalable, with the potential for genome-wide studies that test thousands of genes simultaneously. As sequencing and imaging technologies continue to improve, the researchers anticipate that Perturb-Multi will become even more powerful and accessible to the broader research community. “Our goal is to keep scaling up. We plan to do genome-wide perturbations, study different physiological conditions, and look at different organs,” says Weissman. “That we can now collect so many types of data from so many cells, at speed, is going to be critical for building AI models like virtual cells, and I think it’s going to help us answer previously unsolvable questions about health and disease.”
考点 1:"facet of a multifaceted story" 应译为 "多方面故事的一个侧面" 考点 2:"Perturb-Multimodal (Perturb-Multi)" 应译为 "扰动 - 多模态方法" 考点 3:"multiple types of imaging and sequencing data" 应译为 "多类型成像和测序数据" 考点 4:'living organisms" 应译为 "活体生物" 考点 5:"functional outputs" 应译为 "功能性输出 / 功能结果" 考点 6:"cell properties" 应译为 "细胞特性" 考点 7:"mosaic approach" 应译为 "镶嵌方法 / 镶嵌策略" 考点 8:"variability" 应译为 "变异性 / 差异性" 考点 9:"striking finding" 应译为 "一个引人注目的发现 / 一项惊人的发现" 考点 10:"mechanistic insight" 应译为 "机制性洞见 / 对作用机制的深入理解" 考点 11:"fatty liver disease" 应译为 "脂肪肝疾病" 考点 12:"extracellular matrix — the scaffolding between cells" 应译为 "细胞外基质 — 细胞间的 '支架' " 考点 13:"single-cell gene expression data" 应译为 "单细胞基因表达数据" 考点 14:"bridge multiple disciplines" 应译为 "跨越多个学科 / 在多个学科间架起桥梁" 考点 15:"genomics, imaging, and AI" 应译为 "基因组学、成像与人工智能" 考点 16:"specialized microscopes" 应译为 "专用显微镜" 考点 17:"targeted therapies" 应译为 "靶向疗法" 考点 18:"mechanistic insight" 应译为 "机制性洞见" 考点 19:"cellular stress" 应译为 "细胞应激" 考点 20:"perturbation data" 应译为 "扰动数据"
6aed0
新闻资讯
新闻报道
95
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 2025年6月16日,最高人民检察院发布《未成年人检察工作白皮书(2024)》(简称《白皮书》)显示,2024年,最高检依法核准追诉初中生杀害同学埋尸案等低龄未成年人严重暴力犯罪34人。 2024年,多起低龄未成年实施严重暴力犯罪案件引发广泛关注。其中,2024年3月,河北省邯郸市肥乡区北高镇发生一起3名初中生将同学残忍杀害并将尸体掩埋在蔬菜大棚的恶性案件。最高人民检察院经审查,依法决定对3名犯罪嫌疑人核准追诉。2024年12月30日,邯郸市中级人民法院一审公开宣判该案。3名被告人中,张某某被以故意杀人罪判处无期徒刑,剥夺政治权利终身,李某被以故意杀人罪判处有期徒刑十二年。被告人马某某依法不予刑事处罚。 去年以来,最高检已多次明确,对低龄未成年人严重暴力犯罪,符合核准追诉条件的,要依法核准追诉。最高检还透露,今年将会同相关部门研究出台加强未成年人罪错行为分级干预矫治的意见,研究制定办理低龄未成年人严重暴力犯罪核准追诉案件的意见。 最高检在《白皮书》中表示,检察机关全面准确贯彻宽严相济刑事政策,依法该严则严、当宽则宽。对主观恶性大、犯罪手段残忍的未成年人犯罪依法惩处,决不纵容。同时,对情节较轻、社会危害性较小的犯罪,或者较轻犯罪的初犯、偶犯,依法从宽,最大限度教育挽救,防止再犯。 事件背景 2024年3月10日,河北邯郸肥乡区北高镇张庄村3名初中生霸凌同学,随后将王某某残忍杀害并将尸体掩埋在蔬菜大棚里。3名犯罪嫌疑人心理素质极强,直到调出监控视频,其中的主犯才被指认。找到尸体后发现,受害者已经面目全非。 案件发生后,肥乡区公安机关立即开展侦破工作。2024年3月11日,涉案犯罪嫌疑人被全部抓获,现已依法采取刑事强制措施;3月15日,邯郸市3名涉嫌杀害同学的初中生被刑拘;3月18日,河北邯郸肥乡区警方回应初一学生王某某被杀害案:犯罪嫌疑人为有预谋作案;4月,经最高人民检察院审查,依法决定对犯罪嫌疑人张某某、李某及马某某核准追诉。检察机关审查认为,3名犯罪嫌疑人作案时已满12周岁不满14周岁,故意杀人致被害人死亡,情节恶劣,应当追究刑事责任。 2024年12月30日,河北省邯郸市中级人民法院一审公开宣判被告人张某某、李某、马某某故意杀人一案,对被告人张某某以故意杀人罪判处无期徒刑,剥夺政治权利终身;对被告人李某以故意杀人罪判处有期徒刑十二年;被告人马某某依法不予刑事处罚。邯郸市中级人民法院、邯郸市人民检察院提出司法建议、检察建议,经专门教育指导委员会评估同意,相关公安机关和教育部门依法决定对马某某进行专门矫治教育。
2025年6月16日,最高人民检察院发布《未成年人检察工作白皮书(2024)》(简称《白皮书》)显示,2024年,最高检依法核准追诉初中生杀害同学埋尸案等低龄未成年人严重暴力犯罪34人。 2024年,多起低龄未成年实施严重暴力犯罪案件引发广泛关注。其中,2024年3月,河北省邯郸市肥乡区北高镇发生一起3名初中生将同学残忍杀害并将尸体掩埋在蔬菜大棚的恶性案件。最高人民检察院经审查,依法决定对3名犯罪嫌疑人核准追诉。2024年12月30日,邯郸市中级人民法院一审公开宣判该案。3名被告人中,张某某被以故意杀人罪判处无期徒刑,剥夺政治权利终身,李某被以故意杀人罪判处有期徒刑十二年。被告人马某某依法不予刑事处罚。 去年以来,最高检已多次明确,对低龄未成年人严重暴力犯罪,符合核准追诉条件的,要依法核准追诉。最高检还透露,今年将会同相关部门研究出台加强未成年人罪错行为分级干预矫治的意见,研究制定办理低龄未成年人严重暴力犯罪核准追诉案件的意见。 最高检在《白皮书》中表示,检察机关全面准确贯彻宽严相济刑事政策,依法该严则严、当宽则宽。对主观恶性大、犯罪手段残忍的未成年人犯罪依法惩处,决不纵容。同时,对情节较轻、社会危害性较小的犯罪,或者较轻犯罪的初犯、偶犯,依法从宽,最大限度教育挽救,防止再犯。 事件背景 2024年3月10日,河北邯郸肥乡区北高镇张庄村3名初中生霸凌同学,随后将王某某残忍杀害并将尸体掩埋在蔬菜大棚里。3名犯罪嫌疑人心理素质极强,直到调出监控视频,其中的主犯才被指认。找到尸体后发现,受害者已经面目全非。 案件发生后,肥乡区公安机关立即开展侦破工作。2024年3月11日,涉案犯罪嫌疑人被全部抓获,现已依法采取刑事强制措施;3月15日,邯郸市3名涉嫌杀害同学的初中生被刑拘;3月18日,河北邯郸肥乡区警方回应初一学生王某某被杀害案:犯罪嫌疑人为有预谋作案;4月,经最高人民检察院审查,依法决定对犯罪嫌疑人张某某、李某及马某某核准追诉。检察机关审查认为,3名犯罪嫌疑人作案时已满12周岁不满14周岁,故意杀人致被害人死亡,情节恶劣,应当追究刑事责任。 2024年12月30日,河北省邯郸市中级人民法院一审公开宣判被告人张某某、李某、马某某故意杀人一案,对被告人张某某以故意杀人罪判处无期徒刑,剥夺政治权利终身;对被告人李某以故意杀人罪判处有期徒刑十二年;被告人马某某依法不予刑事处罚。邯郸市中级人民法院、邯郸市人民检察院提出司法建议、检察建议,经专门教育指导委员会评估同意,相关公安机关和教育部门依法决定对马某某进行专门矫治教育。
考点1:“低龄未成年人”推荐译为“ minors of a younger age group; low-age minors” 考点2:“恶性案件”推荐译为“heinous case; vicious case; grave case” 考点3:“罪错行为分级干预矫治”推荐译为“tiered intervention and correction for delinquent and criminal behaviors” 考点4:“宽严相济”推荐译为“ a policy of balancing leniency and severity” 考点5:“主观恶性”推荐译为“subjective malice” 考点6:“心理素质极强”此处形容罪犯,不可译为积极词汇 考点7:“出台……意见”,意见不可用opinion
6bed6
新闻资讯
新闻报道
7
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 黄仁勋越来越焦虑 关注AI产业发展的都能感受到,掌舵英伟达“算力帝国”的黄仁勋,最近越来越焦虑了,开始频繁地提起“中国”和“华为”。 “中国做得太棒了,全球50%的人工智能研究人员都是中国人,你无法阻止他们,也无法阻止他们推进人工智能的发展。如果有人以为,一招就能切断中国发展人工智能的能力,那他绝对是无知的。”黄仁勋近期在台北电脑展上表示。 今年4月,美国政府再度对英伟达中国“特供版”芯片H20发出禁令。公告一出,黄仁勋立刻把皮衣换成西装,飞往中国,这是他3个月里第二次来华。在与政府官员的会面中,黄仁勋多次强调中国市场的重要性,表示希望继续与中国合作。 最新禁令,导致英伟达计提了55亿美元的库存损失,令黄仁勋“深感痛苦”。 事实上,如果仅仅是数十亿美元的损失,对英伟达并不算“伤筋动骨”。这家市值超过3万亿美元的巨头,在AI风口下赚得盆满钵满,2025财年收入达到了1305亿美元,同比增长114%;净利润达到729亿美元,同比增长145%;毛利率达到惊人的75%。 真正让黄仁勋感到焦虑的,是中国自主技术正在美国禁令逼迫下快速突围。在英伟达的关键护城河中,不仅有GPU这种硬件,还有并行计算平台和编程模型CUDA,以及高速互联技术NVLink。在单点竞争上,中国企业可能很难撼动英伟达地位,但系统竞争并不一定。 近期,华为对外详细介绍了昇腾CLoudMatrix 384超节点技术,该技术使用国产昇腾芯片,在算力规模、训推效率和可靠性等关键维度上,全面超越了英伟达最强的NVL72系统。其中的核心在于,华为跳出了单卡算力的竞争,通过计算、存储、网络和架构的协同创新,弥补了硬件和芯片工艺的局限性,最大化发挥了芯片和系统能力。 中国建立了替代英伟达的方案,才是黄仁勋最担心的。这不仅可能导致该公司在中国遭受永久性损失,更可能动摇其“算力帝国”的统治根基。他喊话美国政府:赢得开发者的平台才能最终获胜,出口管制应该强化美国平台,而不是迫使全球一半的AI人才流向竞争对手。 看到中国再度突破,美国政客其实也很着急,但他们显然忽视了黄仁勋放宽管制的呼声。在严格限制英伟达对华出口的同时,美商务部最近还试图在全球封杀华为昇腾芯片,同时逼迫EDA巨头对华断供。但就像黄仁勋说的,这样只会激活中国企业绝处逢生的能力。 就在今年4月,美国对华断供H20之际,华为云率先在芜湖商用了CLoudMatrix 384超节点,随后又在贵安和乌兰察布等地商用,内部人士将其称为“核弹级创新”,今年上半年还将有数万卡规模上线,目标是“彻底终结行业算力焦虑”。而近日华为再度重磅宣布,已经成功在昇腾平台上实现了准万亿MoE模型的全流程训练,集群训练系统的性能上实现了业界领先,进一步验证了国产AI基础设施的自主创新能力。 “四年前,英伟达在中国的市场份额高达95%,如今只有50%。如果我们不在中国竞争,而是让中国开发出新的平台,建立一个丰富的生态系统,并且它们不是美国的,在世界推广人工智能技术的时候,他们的技术和领导力将会传播到世界各地。”不难看出,黄仁勋的焦虑感正越来越重,但美国政客仍在一意孤行,行业对老黄说辞也逐步失去了新鲜感。 现在更令人感兴趣的是,华为是如何在没有先进工艺的情况下,用超节点反超英伟达的?
黄仁勋越来越焦虑 关注AI产业发展的都能感受到,掌舵英伟达“算力帝国”的黄仁勋,最近越来越焦虑了,开始频繁地提起“中国”和“华为”。 “中国做得太棒了,全球50%的人工智能研究人员都是中国人,你无法阻止他们,也无法阻止他们推进人工智能的发展。如果有人以为,一招就能切断中国发展人工智能的能力,那他绝对是无知的。”黄仁勋近期在台北电脑展上表示。 今年4月,美国政府再度对英伟达中国“特供版”芯片H20发出禁令。公告一出,黄仁勋立刻把皮衣换成西装,飞往中国,这是他3个月里第二次来华。在与政府官员的会面中,黄仁勋多次强调中国市场的重要性,表示希望继续与中国合作。 最新禁令,导致英伟达计提了55亿美元的库存损失,令黄仁勋“深感痛苦”。 事实上,如果仅仅是数十亿美元的损失,对英伟达并不算“伤筋动骨”。这家市值超过3万亿美元的巨头,在AI风口下赚得盆满钵满,2025财年收入达到了1305亿美元,同比增长114%;净利润达到729亿美元,同比增长145%;毛利率达到惊人的75%。 真正让黄仁勋感到焦虑的,是中国自主技术正在美国禁令逼迫下快速突围。在英伟达的关键护城河中,不仅有GPU这种硬件,还有并行计算平台和编程模型CUDA,以及高速互联技术NVLink。在单点竞争上,中国企业可能很难撼动英伟达地位,但系统竞争并不一定。 近期,华为对外详细介绍了昇腾CLoudMatrix 384超节点技术,该技术使用国产昇腾芯片,在算力规模、训推效率和可靠性等关键维度上,全面超越了英伟达最强的NVL72系统。其中的核心在于,华为跳出了单卡算力的竞争,通过计算、存储、网络和架构的协同创新,弥补了硬件和芯片工艺的局限性,最大化发挥了芯片和系统能力。 中国建立了替代英伟达的方案,才是黄仁勋最担心的。这不仅可能导致该公司在中国遭受永久性损失,更可能动摇其“算力帝国”的统治根基。他喊话美国政府:赢得开发者的平台才能最终获胜,出口管制应该强化美国平台,而不是迫使全球一半的AI人才流向竞争对手。 看到中国再度突破,美国政客其实也很着急,但他们显然忽视了黄仁勋放宽管制的呼声。在严格限制英伟达对华出口的同时,美商务部最近还试图在全球封杀华为昇腾芯片,同时逼迫EDA巨头对华断供。但就像黄仁勋说的,这样只会激活中国企业绝处逢生的能力。 就在今年4月,美国对华断供H20之际,华为云率先在芜湖商用了CLoudMatrix 384超节点,随后又在贵安和乌兰察布等地商用,内部人士将其称为“核弹级创新”,今年上半年还将有数万卡规模上线,目标是“彻底终结行业算力焦虑”。而近日华为再度重磅宣布,已经成功在昇腾平台上实现了准万亿MoE模型的全流程训练,集群训练系统的性能上实现了业界领先,进一步验证了国产AI基础设施的自主创新能力。 “四年前,英伟达在中国的市场份额高达95%,如今只有50%。如果我们不在中国竞争,而是让中国开发出新的平台,建立一个丰富的生态系统,并且它们不是美国的,在世界推广人工智能技术的时候,他们的技术和领导力将会传播到世界各地。”不难看出,黄仁勋的焦虑感正越来越重,但美国政客仍在一意孤行,行业对老黄说辞也逐步失去了新鲜感。 现在更令人感兴趣的是,华为是如何在没有先进工艺的情况下,用超节点反超英伟达的?
考点1:“算力帝国”可译为“computing power empire; compute empire” 考点2:“ 计提”需译为“write down” 考点3:“护城河”需译为“moat; competitive moat” 考点4:“单点竞争”推荐译为“ single-point competition” 考点5:“ 训推效率”应译为“ training and inference efficiency” 考点6:“ 核弹级创新”译为“ nuclear-level innovation; game-changing innovation” 考点7:“昇腾”必须译为“Ascend”
6cc02
新闻资讯
观点评论
5
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 既往研究中探索了历史街区元素整合的途径。大多数研究未提出营造混合功能的历史街区维持意象一致性的办法。本研究则针对功能混合式历史街区,探讨如何整合不同类型的功能形成的街区意象。本文采用了蒙太奇理论和西蒙的环境经验三分体,选择北京国子监街历史街区为研究对象,分四个步骤分析外来者对该街区的感悟过程,找出那些可以将不同功能的意象整合为完整的历史街区意象的途径。本文得出以下结论:第一, 国子监历史街区内部是文物建筑群、居住区和商业区混合的街区。三类不同的功能区之间有“切断”,不利于该历史街区形成统一的意象. 第二,被试发现四种要素有助于将切断缝合在一起。它们分别是街道两端的大门、商业店铺的招牌、居民区的标语和道路两侧种植的国槐树。基于这两个结论,本文建议对于功能混合的历史街区,可以增加一些具有弥合镜头之间切断作用的要素。 从而让历史街区的意象具有整体性。 历史街区风貌保护是历史街区保护的重要内容。1933年,《雅典宪章》中首次提出保护“历史街区”:“对有历史价值的建筑和街区,均应妥为保存,不可加以破坏”。中国保护历史街区工作开始较晚。1986年中国国务院在公布第二批国家级历史文化名城时,正式提出了保护“历史地段”的概念,古迹比较集中或能较完整地体现出某一历史时期传统风貌和民族地方特色的街区正式列入保护范围。历史性建筑的内外面貌在1987年由国际古迹遗址理事会在华盛顿通过的《保护历史城镇与城区宪章》中被列为历史街区中应该保护的内容中,具体包括体量、形式、建筑风格、材料、建筑装饰等与地段、周围环境的关系,包括与自然和人工环境的关系。注重完整性和和谐性是国外历史街区风貌保护的要点之一,具体方法包括文脉统合 、文脉并置和文脉延续 。 在当下语境中对历史文化街区整体形象的重新构建,能够为保护街区风貌的典型性和完整性找到更有效的手段。林奇提出了可以定义意象的五个元素,分别是:节点、路径、区域、边缘和地标。在历史文化街区中,发展历史“意象”需要更多地关注历史街区的整体形象,历史建筑应与形象背景融合。因而历史街区形象可以被视为由街区风貌的各个维度构成的整体意义或意象。 对于混合功能的历史街区,目前还没有成熟的风格控制方法。既往的研究中探索了历史街区元素整合的途径,大多数研究都强调街区功能和物理要素的一致性。但是遗产需要在主要的社会和文化背景下不断地被重新定义。在现代化与城市化背景下,很多历史文化街区不再是单一的文物保护功能或者居住功能,位于中心城区的历史街区中普遍加入了现代商业、休闲业。之前的方法在混合功能的历史街区中并不适用。人们对功能混杂的历史文化街区的感知具有多种类型,类型之间可能是相互违和的,这造成了街区形象的分裂。关于混合功能的历史街区的形象整合,有学者已经意识到需要将商业服务设施和核心区域的空间配置一起规划,同时适度地控制与分散人流,最大程度保护核心区域。但是尚未有研究提出如何进行统一规划来控制街区风格与形象的整体性。 分析历史街区中外来者整体场所感的形成特征,有助于找到形象整合的途径。 Fakeye和Croupton认为,形象是游客根据自己的印象形成的心理结构。如果想要生成完整立体的街区形象,同时找到破坏形象和利于形象整合的景观要素,需要调查人们流动的地方感。景观意义是由流动性构成的,街道空间也具有明显的流动性和方向性,人们在历史街区中的游览在移动中进行的,相应生成了动态变化的地方感。遗憾的是,之前的研究大多忽视了地方感的动态生成过程,仅调查最后的生成结果。因而,很难准确找到破坏和有利于街区形象整合的景观要素,进行下一步的管控。 因此,本文运用蒙太奇理论,分析外来者能否将被调查街区中不同土地利用的每一个“镜头”整合成一个整体的影像,以及整合的方法是什么,从而为历史街区风貌的保护提出新的建议。 我们使用西蒙的三分体和蒙太奇方法分析了北京文化街区的街区形象。具体分为四个步骤:首先,拍摄国子监街区中每隔约20米的照片,将街区分解成一系列的镜头,根据物理蒙太奇和意义蒙太奇两种蒙太奇逻辑,将镜头拼合成意象,发现其中的切断;然后,在悬隔中找到国子监街的关键主题;接着寻找能够缝合切断的元素;最后,根据这些分析提出一些改善街区形象的建议。 本研究有两个主要结论。第一,国子监历史街区内部是文物建筑群、居住区和商业区混合的街区,建筑形制及听觉嗅觉也不同。不同功能区、不同的建筑形制以及不同的感受之间有“切断”,不利于该历史街区形成统一的意象。第二,我们发现了有助于将切断缝合在一起的要素。它们分别是能够缝合功能区之间切断的商业店铺的招牌和居民区的标语;能够缝合不同建筑形制之间切断街道两段的大门;以及能够消弭噪音带来的切断的国槐。这些要素分别是通过引发外来者的注意或加强外来者与街区之间的强联系来达到消弭切断的效果的。基于这两个结论,本文建议对于功能混合的历史街区,可以增加一些具有弥合镜头之间切断的元素,从而让历史街区的意象具有整体性。而要找到这些具有弥合作用的元素,就需要不断反复与历史街区相遇、悬隔并感悟,在一次次的反思中找到更优路径。
既往研究中探索了历史街区元素整合的途径。大多数研究未提出营造混合功能的历史街区维持意象一致性的办法。本研究则针对功能混合式历史街区,探讨如何整合不同类型的功能形成的街区意象。本文采用了蒙太奇理论和西蒙的环境经验三分体,选择北京国子监街历史街区为研究对象,分四个步骤分析外来者对该街区的感悟过程,找出那些可以将不同功能的意象整合为完整的历史街区意象的途径。本文得出以下结论:第一, 国子监历史街区内部是文物建筑群、居住区和商业区混合的街区。三类不同的功能区之间有“切断”,不利于该历史街区形成统一的意象. 第二,被试发现四种要素有助于将切断缝合在一起。它们分别是街道两端的大门、商业店铺的招牌、居民区的标语和道路两侧种植的国槐树。基于这两个结论,本文建议对于功能混合的历史街区,可以增加一些具有弥合镜头之间切断作用的要素。 从而让历史街区的意象具有整体性。 历史街区风貌保护是历史街区保护的重要内容。1933年,《雅典宪章》中首次提出保护“历史街区”:“对有历史价值的建筑和街区,均应妥为保存,不可加以破坏”。中国保护历史街区工作开始较晚。1986年中国国务院在公布第二批国家级历史文化名城时,正式提出了保护“历史地段”的概念,古迹比较集中或能较完整地体现出某一历史时期传统风貌和民族地方特色的街区正式列入保护范围。历史性建筑的内外面貌在1987年由国际古迹遗址理事会在华盛顿通过的《保护历史城镇与城区宪章》中被列为历史街区中应该保护的内容中,具体包括体量、形式、建筑风格、材料、建筑装饰等与地段、周围环境的关系,包括与自然和人工环境的关系。注重完整性和和谐性是国外历史街区风貌保护的要点之一,具体方法包括文脉统合 、文脉并置和文脉延续 。 在当下语境中对历史文化街区整体形象的重新构建,能够为保护街区风貌的典型性和完整性找到更有效的手段。林奇提出了可以定义意象的五个元素,分别是:节点、路径、区域、边缘和地标。在历史文化街区中,发展历史“意象”需要更多地关注历史街区的整体形象,历史建筑应与形象背景融合。因而历史街区形象可以被视为由街区风貌的各个维度构成的整体意义或意象。 对于混合功能的历史街区,目前还没有成熟的风格控制方法。既往的研究中探索了历史街区元素整合的途径,大多数研究都强调街区功能和物理要素的一致性。但是遗产需要在主要的社会和文化背景下不断地被重新定义。在现代化与城市化背景下,很多历史文化街区不再是单一的文物保护功能或者居住功能,位于中心城区的历史街区中普遍加入了现代商业、休闲业。之前的方法在混合功能的历史街区中并不适用。人们对功能混杂的历史文化街区的感知具有多种类型,类型之间可能是相互违和的,这造成了街区形象的分裂。关于混合功能的历史街区的形象整合,有学者已经意识到需要将商业服务设施和核心区域的空间配置一起规划,同时适度地控制与分散人流,最大程度保护核心区域。但是尚未有研究提出如何进行统一规划来控制街区风格与形象的整体性。 分析历史街区中外来者整体场所感的形成特征,有助于找到形象整合的途径。 Fakeye和Croupton认为,形象是游客根据自己的印象形成的心理结构。如果想要生成完整立体的街区形象,同时找到破坏形象和利于形象整合的景观要素,需要调查人们流动的地方感。景观意义是由流动性构成的,街道空间也具有明显的流动性和方向性,人们在历史街区中的游览在移动中进行的,相应生成了动态变化的地方感。遗憾的是,之前的研究大多忽视了地方感的动态生成过程,仅调查最后的生成结果。因而,很难准确找到破坏和有利于街区形象整合的景观要素,进行下一步的管控。 因此,本文运用蒙太奇理论,分析外来者能否将被调查街区中不同土地利用的每一个“镜头”整合成一个整体的影像,以及整合的方法是什么,从而为历史街区风貌的保护提出新的建议。 我们使用西蒙的三分体和蒙太奇方法分析了北京文化街区的街区形象。具体分为四个步骤:首先,拍摄国子监街区中每隔约20米的照片,将街区分解成一系列的镜头,根据物理蒙太奇和意义蒙太奇两种蒙太奇逻辑,将镜头拼合成意象,发现其中的切断;然后,在悬隔中找到国子监街的关键主题;接着寻找能够缝合切断的元素;最后,根据这些分析提出一些改善街区形象的建议。 本研究有两个主要结论。第一,国子监历史街区内部是文物建筑群、居住区和商业区混合的街区,建筑形制及听觉嗅觉也不同。不同功能区、不同的建筑形制以及不同的感受之间有“切断”,不利于该历史街区形成统一的意象。第二,我们发现了有助于将切断缝合在一起的要素。它们分别是能够缝合功能区之间切断的商业店铺的招牌和居民区的标语;能够缝合不同建筑形制之间切断街道两段的大门;以及能够消弭噪音带来的切断的国槐。这些要素分别是通过引发外来者的注意或加强外来者与街区之间的强联系来达到消弭切断的效果的。基于这两个结论,本文建议对于功能混合的历史街区,可以增加一些具有弥合镜头之间切断的元素,从而让历史街区的意象具有整体性。而要找到这些具有弥合作用的元素,就需要不断反复与历史街区相遇、悬隔并感悟,在一次次的反思中找到更优路径。
考点1:“元素” 推荐翻译为 “element” 考点2:“历史街区” 推荐翻译为 “historic district” 考点3:“文脉并置” 推荐翻译为 “contextual juxtaposition” 考点4:“历史地段” 推荐翻译为 “historic areas”,不可译为“historic sites” 考点5:“悬隔” 必须翻译为 “epoche”,术语译法固定,不可译为其他 考点6:“感悟” 推荐翻译为 “aware” 考点7:“物理蒙太奇和意义蒙太奇”必须译为“physical logic and meaning logic”
6e876
学术论文
人文科学
180
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 2.1 LAG-3功能的早期研究 早期使用LAG-3单克隆抗体的研究表明,在体外阻断LAG-3时,人CD4⁺T细胞克隆的增殖更持久。这种增殖伴随着混合模式的细胞因子产生增强。这些促炎效应仅限于抗原依赖性刺激,在CD8⁺T细胞中未观察到。 这些数据首次提示LAG-3对T细胞功能具有负调控作用,后续使用人细胞的研究证实了这一作用。然而,LAG-3基因敲除动物的构建为探究LAG-3在小鼠模型中T细胞上的作用提供了更精确的手段。这些实验表明,LAG-3在调控CD4⁺和CD8⁺T细胞的体外及体内扩增中发挥作用,从而证实了其负调控因子的身份。 对缺失KIEELE结构域的LAG-3分子的进一步研究显示,该基序在LAG-3的负调控功能中起关键作用,即缺失该结构域的LAG-3分子无法在体外或体内对T细胞功能产生负调控作用。 然而,LAG-3与Ⅱ类MHC分子相互作用的负调控角色存在争议。通过一系列混合淋巴细胞反应,一组研究显示可溶性LAG-3在体外可显著下调人CD4⁺T细胞功能,提示在这些培养条件下,LAG-3与Ⅱ类MHC分子的相互作用具有刺激性。 有趣的是,在人CD8⁺T细胞中未观察到这种对MLR反应的下调,表明CD8⁺T细胞上Ⅱ类分子与LAG-3的相互作用可能在功能上有别于CD4⁺T细胞。这些结果与后续一系列实验存在明显矛盾,在后续实验中,可溶性LAG-3-Ig在体外和体内均被证明可激活树突状细胞。 2.2 LAG-3在调节性T细胞中的作用 通过微阵列分析,我们团队发现,在体内接触自身抗原并呈现调节表型的CD4⁺T细胞上,LAG-3的表达相对上调。在这种自身耐受模型中,我们发现LAG-3阻断抗体可在体内减弱调节性T细胞(Treg)的功能,且用全长而非截短的LAG-3转染抗原特异性CD4⁺T细胞可赋予其体外调节特性。 霍奇金淋巴瘤患者的研究支持这一发现,即患者病情活动时Treg水平升高。体外研究显示,清除LAG-3⁺CD4⁺T细胞可增强肿瘤特异性CD8⁺T细胞的反应性,这与LAG-3在抑制抗肿瘤免疫中的作用一致。 近期研究进一步支持该发现,即来自癌症患者肿瘤部位的LAG-3⁺CD4⁺CD25⁺细胞比LAG-3⁻细胞具有更强的抑制能力。另有研究报道,LAG-3在具有调节功能的FoxP3⁺CD8⁺T细胞亚群中发挥作用,这一新颖发现尤为引人关注,因为调节性CD8⁺T细胞正重新受到重视。 2.3 LAG-3在CD8⁺T细胞中的作用 尽管早期研究对LAG-3在CD8⁺T细胞中的作用提出质疑,但鉴于活化的CD8⁺T细胞中LAG-3的表达量是CD4⁺T细胞的5-8倍,且LAG-3与CD8在活化T细胞中存在相对共定位,这些发现显得颇为奇怪。 利用LAG-3基因敲除动物的研究证实,LAG-3在调控CD8⁺T细胞的稳态增殖以及对超抗原的体内反应中发挥作用。我们通过将抗原特异性CD8⁺T细胞过继转移到携带自身抗原或肿瘤抗原的小鼠体内,进一步证实了这一作用。 在此模型中,LAG-3基因敲除的CD8⁺T细胞显示出增强的增殖和细胞因子产生能力。有趣的是,在过继性T细胞转移前后施用LAG-3阻断抗体也能产生类似的免疫功能增强效果,提示该阻断抗体可能直接作用于CD8⁺T细胞。通过对接受LAG-3基因敲除T细胞过继转移的小鼠施用阻断抗体,这一点得到验证,尽管在该情况下未观察到额外效应。 近期使用CTLA-4阻断抗体的研究发现,其对效应T细胞也存在类似的直接作用,证实免疫检查点阻断在某些情况下可能通过细胞内在机制发挥作用。 关于LAG-3与其他免疫检查点的交叉作用,值得关注的是一项针对慢性病毒感染模型中耗竭CD8⁺T细胞的开创性研究。该研究发现,无功能的CD8⁺T细胞可表达多种检查点分子,部分细胞共表达LAG-3和已知的免疫检查点分子PD-1。在该模型中,同时阻断PD-1和LAG-3比单独阻断任一分子能更有效地改善抗病毒免疫应答。 我们在自身抗原耐受模型中观察到类似表型,即存在一群共表达LAG-3和PD-1的无功能CD8⁺T细胞。这些研究近期已扩展到人类卵巢癌样本,发现在肿瘤抗原特异性CD8⁺T细胞中,有相当比例的细胞共表达LAG-3和PD-1。综上,这些重要研究提示,慢性感染和癌症的免疫治疗可能需要阻断多个免疫检查点。 2.4 LAG-3的作用机制 LAG-3负调控T细胞功能的确切机制尚未完全阐明。如前所述,独特的胞内KIELLE结构域是介导这些效应所必需的。然而,早期对LAG-3的研究发现,在某些患者的血清中存在该分子的可溶性形式,提示LAG-3的 cleavage可能具有一定生理作用。 Vignali团队通过一系列严谨的研究扩展了这些发现,证实LAG-3在细胞表面附近被肿瘤坏死因子α转换酶家族的两个金属蛋白酶(ADAM10和ADAM17)切割。表达不可切割形式的LAG-3会导致T细胞功能出现不可逆缺陷,表明LAG-3的切割是其负调控功能被减弱的主要机制。有趣的是,这些研究未发现切割后的LAG-3片段具有任何作用,这与下文涉及LAG-3-Ig融合蛋白的研究形成鲜明对比。 3 LAG-3在癌症免疫治疗中的应用 3.1 临床前研究 在制备用于生化和功能研究的LAG-3-Ig融合蛋白后不久,研究人员即在小鼠肿瘤模型中对该试剂进行了体内研究。与人类混合淋巴细胞反应的结果以及涉及切割片段的研究不同,可溶性LAG-3-Ig在携带RENCA(肾癌)、MCA205(肉瘤)或TS/A(乳腺癌)肿瘤的小鼠中可介导肿瘤控制和消退。 通过LAG-3转导肿瘤细胞可重复上述发现,提示LAG-3可能通过结合抗原呈递细胞上的Ⅱ类MHC分子并潜在介导其成熟或功能,从而发挥抗肿瘤作用。事实上,利用人单核细胞来源的树突状细胞进行的体外研究证实了这一假设,即LAG-3-Ig可上调共刺激分子的表达并增加树突状细胞中IL-12的产生。 这些表型变化使经LAG-3-Ig成熟的树突状细胞更易介导Th1型应答,表现为应答性T细胞产生的IFN-γ增加。这些结果还提示,LAG-3-Ig可能作为佐剂增强疫苗应答。 事实的确如此,研究显示LAG-3-Ig可显著增强小鼠对可溶性抗原疫苗的CD8⁺T细胞应答以及对颗粒性抗原的体液应答。这种佐剂效应已扩展到癌症疫苗领域,在HER-2/neu转基因小鼠中,LAG-3-Ig与弱DNA疫苗联合施用可预防乳腺癌变。 需要注意的是,这些结果虽令人振奋,但与LAG-3在T细胞增殖和功能中已被充分证实的负调控作用存在一定矛盾。实际上,LAG-3与Ⅱ类分子的相互作用一方面可下调T细胞功能,另一方面却能向表达Ⅱ类分子的树突状细胞传递促炎成熟信号,这一看似矛盾的现象仍有待解释。在此背景下,近期一项研究对LAG-3与Ⅱ类分子结合的促免疫效应提出质疑,认为调节性CD4⁺T细胞上的LAG-3实际上可抑制树突状细胞功能。 3.2 临床研究 LAG-3发现后不久,研究人员注意到部分肾细胞癌(RCC)患者的LAG-3⁺CD4⁺肿瘤浸润淋巴细胞显著扩增。后续研究发现,RCC肿瘤浸润淋巴细胞中LAG-3的表达率为11%-48%,而未检测到明显水平的CTLA-4或检查点分子4-1BB。然而,在这些研究中,LAG-3阻断并未增强CD8⁺T细胞的裂解活性,这提示LAG-3阻断可能在T细胞活化的早期(即起始阶段)更为重要,或反映了使用扩增的人肿瘤浸润淋巴细胞作为试剂存在的技术局限性。 基于LAG-3-Ig在小鼠中的有趣研究结果,该试剂已进入商业化开发并在多项临床试验中接受测试(IMP321,Immutep公司,巴黎)。在首个Ⅰ期试验中,IMP321与标准流感疫苗联合使用,剂量递增。研究未观察到剂量限制性毒性,不良反应轻微。虽未发现体液疫苗应答增强,但在部分受试者中检测到Th1型CD4⁺T细胞应答。 第二项类似试验将IMP321与商用乙肝疫苗联合使用。有趣的是,在较高剂量水平下,单次IMP321处理后即可检测到CD4⁺和CD8⁺T细胞应答。这些结果随后在一项针对肾细胞癌患者的单药剂量递增试验中得到扩展。该药物再次显示出良好的耐受性,且治疗似乎与外周血中CD8⁺(而非CD4⁺)T细胞效应表型的形成相关。 与癌症免疫治疗领域的Ⅰ期试验通常情况一致,本研究未观察到客观应答,但部分患者出现疾病稳定。一项更具创新性的试验将IMP321与紫杉烷类化疗联合用于乳腺癌女性患者。这项单臂试验显示客观缓解率为50%,而历史缓解率约为25%。尽管癌症免疫治疗中的单臂研究需谨慎解读,但多项Ⅱ期试验已在进行中或处于规划阶段。
2.1 LAG-3功能的早期研究 早期使用LAG-3单克隆抗体的研究表明,在体外阻断LAG-3时,人CD4⁺T细胞克隆的增殖更持久。这种增殖伴随着混合模式的细胞因子产生增强。这些促炎效应仅限于抗原依赖性刺激,在CD8⁺T细胞中未观察到。 这些数据首次提示LAG-3对T细胞功能具有负调控作用,后续使用人细胞的研究证实了这一作用。然而,LAG-3基因敲除动物的构建为探究LAG-3在小鼠模型中T细胞上的作用提供了更精确的手段。这些实验表明,LAG-3在调控CD4⁺和CD8⁺T细胞的体外及体内扩增中发挥作用,从而证实了其负调控因子的身份。 对缺失KIEELE结构域的LAG-3分子的进一步研究显示,该基序在LAG-3的负调控功能中起关键作用,即缺失该结构域的LAG-3分子无法在体外或体内对T细胞功能产生负调控作用。 然而,LAG-3与Ⅱ类MHC分子相互作用的负调控角色存在争议。通过一系列混合淋巴细胞反应,一组研究显示可溶性LAG-3在体外可显著下调人CD4⁺T细胞功能,提示在这些培养条件下,LAG-3与Ⅱ类MHC分子的相互作用具有刺激性。 有趣的是,在人CD8⁺T细胞中未观察到这种对MLR反应的下调,表明CD8⁺T细胞上Ⅱ类分子与LAG-3的相互作用可能在功能上有别于CD4⁺T细胞。这些结果与后续一系列实验存在明显矛盾,在后续实验中,可溶性LAG-3-Ig在体外和体内均被证明可激活树突状细胞。 2.2 LAG-3在调节性T细胞中的作用 通过微阵列分析,我们团队发现,在体内接触自身抗原并呈现调节表型的CD4⁺T细胞上,LAG-3的表达相对上调。在这种自身耐受模型中,我们发现LAG-3阻断抗体可在体内减弱调节性T细胞(Treg)的功能,且用全长而非截短的LAG-3转染抗原特异性CD4⁺T细胞可赋予其体外调节特性。 霍奇金淋巴瘤患者的研究支持这一发现,即患者病情活动时Treg水平升高。体外研究显示,清除LAG-3⁺CD4⁺T细胞可增强肿瘤特异性CD8⁺T细胞的反应性,这与LAG-3在抑制抗肿瘤免疫中的作用一致。 近期研究进一步支持该发现,即来自癌症患者肿瘤部位的LAG-3⁺CD4⁺CD25⁺细胞比LAG-3⁻细胞具有更强的抑制能力。另有研究报道,LAG-3在具有调节功能的FoxP3⁺CD8⁺T细胞亚群中发挥作用,这一新颖发现尤为引人关注,因为调节性CD8⁺T细胞正重新受到重视。 2.3 LAG-3在CD8⁺T细胞中的作用 尽管早期研究对LAG-3在CD8⁺T细胞中的作用提出质疑,但鉴于活化的CD8⁺T细胞中LAG-3的表达量是CD4⁺T细胞的5-8倍,且LAG-3与CD8在活化T细胞中存在相对共定位,这些发现显得颇为奇怪。 利用LAG-3基因敲除动物的研究证实,LAG-3在调控CD8⁺T细胞的稳态增殖以及对超抗原的体内反应中发挥作用。我们通过将抗原特异性CD8⁺T细胞过继转移到携带自身抗原或肿瘤抗原的小鼠体内,进一步证实了这一作用。 在此模型中,LAG-3基因敲除的CD8⁺T细胞显示出增强的增殖和细胞因子产生能力。有趣的是,在过继性T细胞转移前后施用LAG-3阻断抗体也能产生类似的免疫功能增强效果,提示该阻断抗体可能直接作用于CD8⁺T细胞。通过对接受LAG-3基因敲除T细胞过继转移的小鼠施用阻断抗体,这一点得到验证,尽管在该情况下未观察到额外效应。 近期使用CTLA-4阻断抗体的研究发现,其对效应T细胞也存在类似的直接作用,证实免疫检查点阻断在某些情况下可能通过细胞内在机制发挥作用。 关于LAG-3与其他免疫检查点的交叉作用,值得关注的是一项针对慢性病毒感染模型中耗竭CD8⁺T细胞的开创性研究。该研究发现,无功能的CD8⁺T细胞可表达多种检查点分子,部分细胞共表达LAG-3和已知的免疫检查点分子PD-1。在该模型中,同时阻断PD-1和LAG-3比单独阻断任一分子能更有效地改善抗病毒免疫应答。 我们在自身抗原耐受模型中观察到类似表型,即存在一群共表达LAG-3和PD-1的无功能CD8⁺T细胞。这些研究近期已扩展到人类卵巢癌样本,发现在肿瘤抗原特异性CD8⁺T细胞中,有相当比例的细胞共表达LAG-3和PD-1。综上,这些重要研究提示,慢性感染和癌症的免疫治疗可能需要阻断多个免疫检查点。 2.4 LAG-3的作用机制 LAG-3负调控T细胞功能的确切机制尚未完全阐明。如前所述,独特的胞内KIELLE结构域是介导这些效应所必需的。然而,早期对LAG-3的研究发现,在某些患者的血清中存在该分子的可溶性形式,提示LAG-3的 cleavage可能具有一定生理作用。 Vignali团队通过一系列严谨的研究扩展了这些发现,证实LAG-3在细胞表面附近被肿瘤坏死因子α转换酶家族的两个金属蛋白酶(ADAM10和ADAM17)切割。表达不可切割形式的LAG-3会导致T细胞功能出现不可逆缺陷,表明LAG-3的切割是其负调控功能被减弱的主要机制。有趣的是,这些研究未发现切割后的LAG-3片段具有任何作用,这与下文涉及LAG-3-Ig融合蛋白的研究形成鲜明对比。 3 LAG-3在癌症免疫治疗中的应用 3.1 临床前研究 在制备用于生化和功能研究的LAG-3-Ig融合蛋白后不久,研究人员即在小鼠肿瘤模型中对该试剂进行了体内研究。与人类混合淋巴细胞反应的结果以及涉及切割片段的研究不同,可溶性LAG-3-Ig在携带RENCA(肾癌)、MCA205(肉瘤)或TS/A(乳腺癌)肿瘤的小鼠中可介导肿瘤控制和消退。 通过LAG-3转导肿瘤细胞可重复上述发现,提示LAG-3可能通过结合抗原呈递细胞上的Ⅱ类MHC分子并潜在介导其成熟或功能,从而发挥抗肿瘤作用。事实上,利用人单核细胞来源的树突状细胞进行的体外研究证实了这一假设,即LAG-3-Ig可上调共刺激分子的表达并增加树突状细胞中IL-12的产生。 这些表型变化使经LAG-3-Ig成熟的树突状细胞更易介导Th1型应答,表现为应答性T细胞产生的IFN-γ增加。这些结果还提示,LAG-3-Ig可能作为佐剂增强疫苗应答。 事实的确如此,研究显示LAG-3-Ig可显著增强小鼠对可溶性抗原疫苗的CD8⁺T细胞应答以及对颗粒性抗原的体液应答。这种佐剂效应已扩展到癌症疫苗领域,在HER-2/neu转基因小鼠中,LAG-3-Ig与弱DNA疫苗联合施用可预防乳腺癌变。 需要注意的是,这些结果虽令人振奋,但与LAG-3在T细胞增殖和功能中已被充分证实的负调控作用存在一定矛盾。实际上,LAG-3与Ⅱ类分子的相互作用一方面可下调T细胞功能,另一方面却能向表达Ⅱ类分子的树突状细胞传递促炎成熟信号,这一看似矛盾的现象仍有待解释。在此背景下,近期一项研究对LAG-3与Ⅱ类分子结合的促免疫效应提出质疑,认为调节性CD4⁺T细胞上的LAG-3实际上可抑制树突状细胞功能。 3.2 临床研究 LAG-3发现后不久,研究人员注意到部分肾细胞癌(RCC)患者的LAG-3⁺CD4⁺肿瘤浸润淋巴细胞显著扩增。后续研究发现,RCC肿瘤浸润淋巴细胞中LAG-3的表达率为11%-48%,而未检测到明显水平的CTLA-4或检查点分子4-1BB。然而,在这些研究中,LAG-3阻断并未增强CD8⁺T细胞的裂解活性,这提示LAG-3阻断可能在T细胞活化的早期(即起始阶段)更为重要,或反映了使用扩增的人肿瘤浸润淋巴细胞作为试剂存在的技术局限性。 基于LAG-3-Ig在小鼠中的有趣研究结果,该试剂已进入商业化开发并在多项临床试验中接受测试(IMP321,Immutep公司,巴黎)。在首个Ⅰ期试验中,IMP321与标准流感疫苗联合使用,剂量递增。研究未观察到剂量限制性毒性,不良反应轻微。虽未发现体液疫苗应答增强,但在部分受试者中检测到Th1型CD4⁺T细胞应答。 第二项类似试验将IMP321与商用乙肝疫苗联合使用。有趣的是,在较高剂量水平下,单次IMP321处理后即可检测到CD4⁺和CD8⁺T细胞应答。这些结果随后在一项针对肾细胞癌患者的单药剂量递增试验中得到扩展。该药物再次显示出良好的耐受性,且治疗似乎与外周血中CD8⁺(而非CD4⁺)T细胞效应表型的形成相关。 与癌症免疫治疗领域的Ⅰ期试验通常情况一致,本研究未观察到客观应答,但部分患者出现疾病稳定。一项更具创新性的试验将IMP321与紫杉烷类化疗联合用于乳腺癌女性患者。这项单臂试验显示客观缓解率为50%,而历史缓解率约为25%。尽管癌症免疫治疗中的单臂研究需谨慎解读,但多项Ⅱ期试验已在进行中或处于规划阶段。
考点1:“无功能的 CD8⁺T 细胞” 中的 “无功能”,不可以翻译为“non-functional”,必须译为“dysfunctional”或“exhausted” 考点2:“裂解活性” 必须译为“lytic activity (against target cells)”或“cytolytic activity”。 考点3:文中的“不良反应轻微” 描述的是临床试验中的不良事件,不可以翻译为“adverse reactions” 。 考点4:“紫杉烷类化疗” 不能直接翻译为“taxane chemotherapy” ,必须译为“taxane-based chemotherapy” 考点5:“调节性 T 细胞(Treg)” 推荐译为 “regulatory T cell (Treg)” 考点6:“Ⅱ 类 MHC 分子” 推荐译为 “Class II MHC molecule” 考点7:“混合淋巴细胞反应(MLR)” 必须译为 “mixed lymphocyte reaction (MLR)” 考点8:“KIEELE 结构域” 必须译为 “KIEELE domain”不可误译为 “KIEELE motif”,以免引起生物学概念错误。 考点9:“树突状细胞” 推荐译为 “dendritic cell” 考点10:“肿瘤浸润淋巴细胞(TILs)” 推荐译为 “tumor-infiltrating lymphocytes (TILs)” 考点11:“过继转移” 推荐译为 “adoptive transfer” 考点12:“Th1 型应答” 推荐译为 “Th1 response” 考点13:“负调控” 不可翻译为 “negative regulatory effect”。 考点14:混合淋巴细胞反应的缩写是 “MLR”,在专有名词第一次出现之后,后续如果用缩略词,需要在第一次出现该专有名词之后,使用缩写表述。 考点15:“单核细胞来源的树突状细胞”推荐译为“monocyte-derived dendritic cells” 考点16:“转基因小鼠”推荐译为“transgenic mice” 考点17:“免疫检查点阻断”推荐译为“immune checkpoint blockade” 考点18:“促炎成熟信号”推荐译为“pro-inflammatory maturation signal” 考点19:“单臂试验”推荐译为“single-arm trial” 考点20:“促炎效应”强调“促进炎症的生物学效应”,在免疫学中有固定搭配:“pro-inflammatory” 考点21:“调节表型”中,“表型”在生物学中对应“phenotype”。不可翻译为 “type”。 考点22:“抗肿瘤免疫”可译为“antitumor immunity” 考点23:“剂量递增”固定译法为“dose-escalation”。不可翻译为 “increasing dose”。 考点24:“细胞因子产生能力”在免疫学文献中常直接表达为 “cytokine production”,不推荐译为 “cytokine production capacity”,以保持简洁和学术规范。 考点25:“反映出使用……的技术限制” 推荐译为 “reflecting the technical limitations associated with using...” 或 “reflecting technical limitations inherent in the use of...”,不推荐直译为 “reflecting technical limitations of using...”。 考点26:表达“显得奇怪”时,学术写作中应译为 “unexpected” 或 “surprising”,避免使用过于口语化的 “strange” 或 “quite strange”。
72efc
学术论文
自然科学
115
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 近年来,自动驾驶领域革新迭起,世界模型研究在决策推理与安全保障中愈发彰显关键价值,其核心要义在于通过对环境动态结构及隐含规律进行层级化解构与重构,构建多层次深度表征体系,从而赋予系统跨场景环境预测推理能力,为高维动作空间中的可解释规划控制提供核心支撑。传统依托几何约束与物理规则的导航算法,高度依赖高精地图与经典图搜索技术生成全局最优路径,此类方法虽能在受控道路网络中完成路径可行性验证,却因缺乏对环境时变要素的高效自适应机制,难以应对多交通参与主体行为的非线性耦合及突发临界事件。 随着深度学习理论与算力资源的突破性发展,学术界率先探索神经网络与逆向强化学习范式的融合路径,实现决策与世界模型结构的同步训练。此类框架通过采集海量仿真或实车轨迹数据反演潜在奖励函数,并借助时序隐变量生成模型学习状态转移概率密度函数。其中最具代表性的研究方案,采用条件变分自编码器对观测序列进行端到端重建,同时融合对抗生成模型提升样本空间覆盖度,结合信息瓶颈约束强化模型抽象表征的效率与鲁棒性。与此同时,多模态感知融合技术步入成熟阶段,其演进路径从单一摄像头与激光雷达的信号级拼接,逐步拓展至整合毫米波雷达、超声波传感器及高精度实时定位信息,并通过自注意力机制在各传感通道间实现动态加权,进而在极端光照、复杂气象及高密度交通场景下保持鲁棒感知与低时延响应性能。 在此基础上,研究社区进一步推动世界模型与可微分规划控制算法的深度耦合,创新性提出端到端微分物理仿真管道。该架构通过集成微分动力学求解器,在前向推断过程中融合运动学约束与碰撞检测模块,并以内嵌梯度下降算法优化感知、预测、控制子模块的联合更新机制,实现对不确定性场景的可解释联合优化。这一创新思路打破传统黑箱策略网络的局限,将决策过程映射为显式优化问题,赋予系统高度可验证性与多场景泛化能力。 伴随工业界对安全测试与场景覆盖率需求的激增,CARLA、LGSVL 等大规模联合仿真平台及自研数字孪生环境应运而生。这些平台支持多源高精地图全局叠加、物理光照光谱模型与仿真行为体部署,并可加载交通流热力学平衡模型,实现对城市级道路网络多时空维度的全景式仿真推演。部分顶尖研究团队更将自监督预训练与对比学习范式引入世界模型训练流程,利用动态伪标签生成与编码器蒸馏技术,在低标注成本下实现高精度特征提取,并通过少样本自适应算法加速新场景迁移适配。 在架构设计层面,当前学术与产业界的主流技术路线可划分为三大阵营:纯强化学习世界模型、混合结构化神经符号模型及纯端到端深度网络。其中,纯强化学习范式依托隐状态嵌入与价值迭代算法,实现对高维连续动作空间的无模型控制,但其在样本效率与收敛速度方面存在固有瓶颈;混合结构化模型通过物理知识图神经网络与符号推理模块的协同,实现先验知识与数据驱动的有机融合,可解释性表现优异,却需设计者投入大量领域专业知识进行模型构建;纯端到端深度网络阵营聚焦大规模 Transformer 架构与序列决策流程的结合,虽具备空前的表达能力,却需承担海量标注数据与算力成本,且在安全性验证与实时性控制方面面临严峻挑战。 当前研究焦点集中于如何在保障微秒级决策时延与动态场景实时响应的前提下,提升对罕见关键场景的前瞻性风险评估能力。学者们通过引入风险敏感度损失函数与信息熵驱动探索策略,使模型在训练过程中自动聚焦潜在高危状态;多智能体协同共享世界模型成为另一研究热点,相关团队正借助联邦学习与在线知识蒸馏框架,在保障数据隐私的前提下实现跨车队表征共享,并结合分布式边缘计算将联合推理模块部署至车载终端,以降低通信时延并提升系统鲁棒性。 工业界领军企业已在二级高速及城市复杂路口部署多款世界模型驱动系统,在百万公里实车测试中实现平均决策时延低于 25 毫秒、危急事件预警误报率低于 1%,同时通过事件驱动最优控制策略与多模态对齐算法,在拥堵场景下最大限度降低能耗与碳排放,并支持持续在线升级迭代。 尽管上述突破为安全驾驶与可解释决策奠定了坚实基础,但世界模型仍面临长时序预测误差累积、形式化验证工具匮乏、法律伦理边界未明及计算资源与标签成本制约等多重挑战。未来,多学科交叉的形式化验证方法、可解释性增强技术及自治伦理评估体系的构建,将是突破当前发展瓶颈的关键方向。
近年来,自动驾驶领域革新迭起,世界模型研究在决策推理与安全保障中愈发彰显关键价值,其核心要义在于通过对环境动态结构及隐含规律进行层级化解构与重构,构建多层次深度表征体系,从而赋予系统跨场景环境预测推理能力,为高维动作空间中的可解释规划控制提供核心支撑。传统依托几何约束与物理规则的导航算法,高度依赖高精地图与经典图搜索技术生成全局最优路径,此类方法虽能在受控道路网络中完成路径可行性验证,却因缺乏对环境时变要素的高效自适应机制,难以应对多交通参与主体行为的非线性耦合及突发临界事件。 随着深度学习理论与算力资源的突破性发展,学术界率先探索神经网络与逆向强化学习范式的融合路径,实现决策与世界模型结构的同步训练。此类框架通过采集海量仿真或实车轨迹数据反演潜在奖励函数,并借助时序隐变量生成模型学习状态转移概率密度函数。其中最具代表性的研究方案,采用条件变分自编码器对观测序列进行端到端重建,同时融合对抗生成模型提升样本空间覆盖度,结合信息瓶颈约束强化模型抽象表征的效率与鲁棒性。与此同时,多模态感知融合技术步入成熟阶段,其演进路径从单一摄像头与激光雷达的信号级拼接,逐步拓展至整合毫米波雷达、超声波传感器及高精度实时定位信息,并通过自注意力机制在各传感通道间实现动态加权,进而在极端光照、复杂气象及高密度交通场景下保持鲁棒感知与低时延响应性能。 在此基础上,研究社区进一步推动世界模型与可微分规划控制算法的深度耦合,创新性提出端到端微分物理仿真管道。该架构通过集成微分动力学求解器,在前向推断过程中融合运动学约束与碰撞检测模块,并以内嵌梯度下降算法优化感知、预测、控制子模块的联合更新机制,实现对不确定性场景的可解释联合优化。这一创新思路打破传统黑箱策略网络的局限,将决策过程映射为显式优化问题,赋予系统高度可验证性与多场景泛化能力。 伴随工业界对安全测试与场景覆盖率需求的激增,CARLA、LGSVL 等大规模联合仿真平台及自研数字孪生环境应运而生。这些平台支持多源高精地图全局叠加、物理光照光谱模型与仿真行为体部署,并可加载交通流热力学平衡模型,实现对城市级道路网络多时空维度的全景式仿真推演。部分顶尖研究团队更将自监督预训练与对比学习范式引入世界模型训练流程,利用动态伪标签生成与编码器蒸馏技术,在低标注成本下实现高精度特征提取,并通过少样本自适应算法加速新场景迁移适配。 在架构设计层面,当前学术与产业界的主流技术路线可划分为三大阵营:纯强化学习世界模型、混合结构化神经符号模型及纯端到端深度网络。其中,纯强化学习范式依托隐状态嵌入与价值迭代算法,实现对高维连续动作空间的无模型控制,但其在样本效率与收敛速度方面存在固有瓶颈;混合结构化模型通过物理知识图神经网络与符号推理模块的协同,实现先验知识与数据驱动的有机融合,可解释性表现优异,却需设计者投入大量领域专业知识进行模型构建;纯端到端深度网络阵营聚焦大规模 Transformer 架构与序列决策流程的结合,虽具备空前的表达能力,却需承担海量标注数据与算力成本,且在安全性验证与实时性控制方面面临严峻挑战。 当前研究焦点集中于如何在保障微秒级决策时延与动态场景实时响应的前提下,提升对罕见关键场景的前瞻性风险评估能力。学者们通过引入风险敏感度损失函数与信息熵驱动探索策略,使模型在训练过程中自动聚焦潜在高危状态;多智能体协同共享世界模型成为另一研究热点,相关团队正借助联邦学习与在线知识蒸馏框架,在保障数据隐私的前提下实现跨车队表征共享,并结合分布式边缘计算将联合推理模块部署至车载终端,以降低通信时延并提升系统鲁棒性。 工业界领军企业已在二级高速及城市复杂路口部署多款世界模型驱动系统,在百万公里实车测试中实现平均决策时延低于 25 毫秒、危急事件预警误报率低于 1%,同时通过事件驱动最优控制策略与多模态对齐算法,在拥堵场景下最大限度降低能耗与碳排放,并支持持续在线升级迭代。 尽管上述突破为安全驾驶与可解释决策奠定了坚实基础,但世界模型仍面临长时序预测误差累积、形式化验证工具匮乏、法律伦理边界未明及计算资源与标签成本制约等多重挑战。未来,多学科交叉的形式化验证方法、可解释性增强技术及自治伦理评估体系的构建,将是突破当前发展瓶颈的关键方向。
考点1 :“环境动态结构及隐含规律” 推荐译为 environmental dynamics structure and implicit laws 考点2 :“多层次深度表征体系” 推荐译为 multi‑level deep representation framework 考点3 :“多模态感知融合技术” 必须译为 multimodal perception fusion technology 考点4 :“少样本自适应算法” 必须译为 few‑shot adaptation algorithm
74ae3
学术论文
应用学科
149
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Section One------ How to teach reading I. Why teach reading There are many reasons why getting students to read English texts is an important part of the teacher’s job. In the first place, many of them want to be able to read texts in English either for their careers, for study purposes or simply for pleasure. Anything we can do to make reading easier for them must be a good idea. Reading texts provide good models for English writing, provide opportunities to study language vocabulary, grammar, punctuation, and the way to construct sentences, paragraphs and texts. Lastly, good reading texts can introduce interesting topics, stimulate discussion, excite imaginative responses and be the springboard for well-rounded, fascinating lessons. The last but not the least, students must read widely because only a fraction of knowledge about the world can come from other experiences in their short lives. II. What kind of reading should students do? When the teachers give reading class to students, they should notice a balance----a balance to be struck between real English on the one hand and the students’ capabilities and interests on the other. There is some authentic written material which beginner students can understand to some degree: menus, timetables, signs and basic instructions, for example, and, where appropriate, teachers can use these. But for longer prose, teachers can offer their students texts, which, while being like English, are nevertheless written or adapted especially for their level. Anyway, the materials to be read should be interesting and meaningful. Teachers should become better acquainted with books written specially for teenagers and dealing with their problems. III. What are the principles behind the teaching of reading? i) Permit Students To Read No one has learned to swim by practicing the skills of backstrokes, flutter kicks or treading water while staying on the edge of the swimming pool. Yet, in the teaching of reading teachers often do just that. Rather than let the students into “the water”, teachers keep them in skills books learning rules about letters, syllables or definitions of words rather than letting them into the book itself, permitting them to be immersed in the language which comes from the authors as the readers try to reconstruct the written message. ii) Encourage students to respond to the content of a reading text, not just to the language Of course, it is important to study reading texts for the way they use language, how many paragraphs they contain and how many times they use relative clauses. But the meaning, the message of the text, is much more important. Teachers should help students understand that the main reason to read is for them. They have to have their own purpose to read and reading must make sense, they have to find ways of doing something about it. They should be encouraged either to reread or to continue reading to gain meaning. But they must realize that the meaning is not in the teacher, but in the interaction between the reader and author. Students should be encouraged to ask themselves repeatedly, “Does this make sense to me?” Students should be encouraged to reject and to be intolerant of reading materials that do not make sense. iii) Encourage students to guess or predict Readers’ guesses or predictions are based on the cumulative information and syntactic structure they have been learning as they have been reading. Therefore, their guesses are more often than not appropriate to the materials. Students have to realize that risk taking in reading is appropriate; that using context to decide what words mean is a proficient reading strategy and that they have the language sense to make appropriate guesses which can fit both the grammatical and semantic sense of what they are reading. iv) Match the task to the topic Once a decision has been taken about what kind of reading text the students are going to read, teachers need to choose good reading tasks—the right kind of questions and useful puzzles, etc. Asking boring and inappropriate questions can undermine the most interesting text; the most commonplace passage can be made really exciting with imaginative and challenging tasks. Working in groups, the English teacher and students take turns asking each other questions following the reading. The teacher may ask, “ What is the significance of the character’s age?” These questions require inferences based on details from the reading text. Section Two------How to teach writing (Developing correctness in students’ writing) “Students learn to write by writing, and they learn to write correctly by writing, revising, and proofreading their own work”---with some help or direction from the teacher when it is necessary. They do not learn to write correctly by studying about writing or doing isolated workbook exercises unrelated to their own writing. So, the most important technique a teacher can use to guide students toward grammatically correct writing is to let them write, let them write things related to their own experiences. There is no limit to the kinds of text the teacher can ask students to write. Teachers’ decisions, though, should based on how much language the students know, what their interests are. “Do I read a paper and ignore all punctuation, what good is that for students? We spend hours at night with papers---I’m not sure the students get as much from it as the time I spend on it.” These comments by senior high school English teachers discussing the process of marking student papers reflect the dissatisfaction and frustration of many teachers over the problem of dealing with the errors in student writing-----the obvious mistakes in spelling, punctuation----Traditionally, teachers have worked to correct errors in two ways: by teaching grammatically correctness through exercise in grammar texts; by pointing out all errors when making student papers. Most students find it very dispiriting if they get a piece of written work back and it is covered in red ink, underlings and crossing-out. It is a powerful visual statement of the fact that their written English is terrible. Of course, some pieces of written work are completely full of mistakes, but even in these cases, the teacher has to achieve a balance between being accurate and truthful on the one hand and treating students sensitively and sympathetically on the other.
Section One------ How to teach reading I. Why teach reading There are many reasons why getting students to read English texts is an important part of the teacher’s job. In the first place, many of them want to be able to read texts in English either for their careers, for study purposes or simply for pleasure. Anything we can do to make reading easier for them must be a good idea. Reading texts provide good models for English writing, provide opportunities to study language vocabulary, grammar, punctuation, and the way to construct sentences, paragraphs and texts. Lastly, good reading texts can introduce interesting topics, stimulate discussion, excite imaginative responses and be the springboard for well-rounded, fascinating lessons. The last but not the least, students must read widely because only a fraction of knowledge about the world can come from other experiences in their short lives. II. What kind of reading should students do? When the teachers give reading class to students, they should notice a balance----a balance to be struck between real English on the one hand and the students’ capabilities and interests on the other. There is some authentic written material which beginner students can understand to some degree: menus, timetables, signs and basic instructions, for example, and, where appropriate, teachers can use these. But for longer prose, teachers can offer their students texts, which, while being like English, are nevertheless written or adapted especially for their level. Anyway, the materials to be read should be interesting and meaningful. Teachers should become better acquainted with books written specially for teenagers and dealing with their problems. III. What are the principles behind the teaching of reading? i) Permit Students To Read No one has learned to swim by practicing the skills of backstrokes, flutter kicks or treading water while staying on the edge of the swimming pool. Yet, in the teaching of reading teachers often do just that. Rather than let the students into “the water”, teachers keep them in skills books learning rules about letters, syllables or definitions of words rather than letting them into the book itself, permitting them to be immersed in the language which comes from the authors as the readers try to reconstruct the written message. ii) Encourage students to respond to the content of a reading text, not just to the language Of course, it is important to study reading texts for the way they use language, how many paragraphs they contain and how many times they use relative clauses. But the meaning, the message of the text, is much more important. Teachers should help students understand that the main reason to read is for them. They have to have their own purpose to read and reading must make sense, they have to find ways of doing something about it. They should be encouraged either to reread or to continue reading to gain meaning. But they must realize that the meaning is not in the teacher, but in the interaction between the reader and author. Students should be encouraged to ask themselves repeatedly, “Does this make sense to me?” Students should be encouraged to reject and to be intolerant of reading materials that do not make sense. iii) Encourage students to guess or predict Readers’ guesses or predictions are based on the cumulative information and syntactic structure they have been learning as they have been reading. Therefore, their guesses are more often than not appropriate to the materials. Students have to realize that risk taking in reading is appropriate; that using context to decide what words mean is a proficient reading strategy and that they have the language sense to make appropriate guesses which can fit both the grammatical and semantic sense of what they are reading. iv) Match the task to the topic Once a decision has been taken about what kind of reading text the students are going to read, teachers need to choose good reading tasks—the right kind of questions and useful puzzles, etc. Asking boring and inappropriate questions can undermine the most interesting text; the most commonplace passage can be made really exciting with imaginative and challenging tasks. Working in groups, the English teacher and students take turns asking each other questions following the reading. The teacher may ask, “ What is the significance of the character’s age?” These questions require inferences based on details from the reading text. Section Two------How to teach writing (Developing correctness in students’ writing) “Students learn to write by writing, and they learn to write correctly by writing, revising, and proofreading their own work”---with some help or direction from the teacher when it is necessary. They do not learn to write correctly by studying about writing or doing isolated workbook exercises unrelated to their own writing. So, the most important technique a teacher can use to guide students toward grammatically correct writing is to let them write, let them write things related to their own experiences. There is no limit to the kinds of text the teacher can ask students to write. Teachers’ decisions, though, should based on how much language the students know, what their interests are. “Do I read a paper and ignore all punctuation, what good is that for students? We spend hours at night with papers---I’m not sure the students get as much from it as the time I spend on it.” These comments by senior high school English teachers discussing the process of marking student papers reflect the dissatisfaction and frustration of many teachers over the problem of dealing with the errors in student writing-----the obvious mistakes in spelling, punctuation----Traditionally, teachers have worked to correct errors in two ways: by teaching grammatically correctness through exercise in grammar texts; by pointing out all errors when making student papers. Most students find it very dispiriting if they get a piece of written work back and it is covered in red ink, underlings and crossing-out. It is a powerful visual statement of the fact that their written English is terrible. Of course, some pieces of written work are completely full of mistakes, but even in these cases, the teacher has to achieve a balance between being accurate and truthful on the one hand and treating students sensitively and sympathetically on the other.
考点1:“reading texts provide good models for English writing”推荐译为“阅读材料为英语写作提供良好范式”,“good models”不可译为“写作模板”或“写作模型”。 考点2:“permit students to read”推荐译为“让学生真正进行实际阅读”,不可直译为“允许学生读书”或“许可学生阅读”。 考点3:“the meaning is not in the teacher”推荐译为“意义不在教师本身”,不可理解为“老师不讲意义”或“教师无意义”。 考点4:“balance between being accurate and being sympathetic”推荐译为“在准确指出错误与富有同理心之间取得平衡”
759ac
垂类场景
教育
25
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: FOR DECADES Sweden was seen as the epitome of sexual freedom, so much so that President Dwight Eisenhower fulminated in 1960 that its people tended towards “sin, nudity, drunkenness”. In 1971 it followed Denmark to become the second country in the world to legalise all forms of pornography. Yet Sweden has been altogether more prudish when it comes to prostitution, having originated the so-called Nordic Model in 1999, which criminalised the purchase of sex, but not its sale, with the intention of reducing demand while protecting vulnerable women. This model has since spread widely. In the past decade, France, Ireland, Israel and the American state of Maine have all adopted it; Scotland is considering it. Now Sweden is trying to apply its real-world Nordic model to the digital world.On July 1st a new Swedish law will come into force that criminalises paying for live porn on sites like OnlyFans, the platform best known for its adult content, but not those who perform the online sex acts. Those breaking the new law face a penalty of up to a year in prison. The new law comes as sex workers increasingly embrace technology and move online. Many of those selling physical sex have already transitioned from soliciting on street corners to advertising on the internet. More recently the sale of virtual sex has gained momentum. Buying sex and sexual content has also become easier-and more accepted-than ever. Fully 14% of young Americans say they would consider selling content on OnlyFans; 16% say they would become a "sugar baby", selling companionship (and very often sex) to older men. In May 2024 one in ten British adults who were online visited Chaturbate, a live webcam-sex site. And the boundaries of what people consider to be sex work are becoming blurry, notes Teela Sanders of the University of Leicester. In Sweden, 8% of girls aged 15-19 say they have sent sexual content or met someone for sex in exchange for money often via the social-media app Snapchat. On the gay dating app Grindr, it is common to trade sex for gifts or drugs. In an age where sex and relationships are often more overtly transactional , many in Generation Z do not see what they are doing as "sex work" at all (unlike, for example, work in brothels, which is usually perceived as more exploitative). Only 56% of Britons aged 18-25 say that"sugaring" counts as sex work, compared with 70% of over-65s. These changes raise two important questions. Does the Swedish model work in the real world? And should it also apply to virtual sex work? Supporters of the Nordic model in relation to old-school prostitution argue that the purchase of sex is always exploitative, and that restrictions are needed not only to protect the vulnerable but also to crush demand. In this they hark back to abolitionist views of the 1920s, when the League of Nations sent undercover agents to infiltrate brothels, where they found examples of foreign women in debt bondage. Denying sex workers agency over their own bodies, they declared that many were "mentally abnormal" and of "poor heredity and poor environment" Though attitudes have softened since, some of the stigma and denial of agency persist in the Nordic model, which is based on the idea that many prostitutes go into sex work because of childhood abuse or poverty rather than free choice, and that they should be "rehabilitated". To be sure, there are many sex workers like Samuel Vahlund, who turned to prostitution after childhood bullying and a traumatic sexual assault destroyed his self-esteem, and because he needed money for drugs. At the time, he believed he was consenting, but would dissociate during sex with other men. “I hated myself,” he says. And similar trends appear to be the case for many of those entering online sex work. Meghan Donevan, a researcher at Talita, a Swedish charity, interviewed 120 people who had appeared in “commercial pornography”. Of those, 88% had been sexually abused as children and 69% had attempted suicide. Virtual sex work also creates the possibility for other sorts of harm. In thousands of webcam “studios” across Colombia, women work 12-hour shifts streaming from cramped, often filthy cubicles, keeping only a fraction of their earnings. “Doxxing”, whereby personal information is published on the internet, is common. “I know a lot of people who prefer old-school sex work because it’s safer,” says one Italian sex worker. Sweden’s government argues that since virtual sex work also carries the risk of harm, it should get the same legal treatment as the sale of actual sex. Moreover, says Nina Larson, Sweden’s minister for gender equality, “digital prostitution” can be a “gateway” to selling sex in person. Does the argument have merit? Supporters of the existing Nordic model tout a fall in the share of Swedish men who say they have ever paid for sex, from 14% in 1996 to 9% in 2017. They also point to falling street prostitution as proof of its success. But that may well be because sex work has moved underground or out of the country. The government reckons that 80% of Swedish men who pay for sex do so when they are abroad. Janna Davidson, Sweden’s national rapporteur on human trafficking and prostitution, argues that the most important effect of the Nordic model has been normative: Swedes are far more likely than their rich-world peers to say prostitution can “never be justified”. Yet there is little firm evidence that the Nordic model is achieving its main goals of protecting vulnerable sex workers or of putting a lasting dent in demand, especially since a growing proportion of men suspected of buying sex are under 30 Worse, the policies in force since 1999 may have distracted authorities from reducing the genuinely harmful aspects of the sex trade. This is because police often find it easier to arrest johns than to pursue pimps and traffickers. Prosecutions for buying sex have risen in Sweden, but there were no convictions for trafficking in 2024. Far from protecting vulnerable people, the existing laws may well be harming them. The stigmatisation of sex work means prostitutes and their clients are less likely to seek treatment for sexually transmitted infections. Migrants—who may make up 70-80% of those selling sex—are particularly vulnerable, because even though the sale of sex is not criminalised, it can be grounds for deportation. Paulina Bolton, who co-ordinates investigations into prostitution and human trafficking at the Swedish Gender Equality Agency, admits that support for victims is lacking. The European Sex Workers Rights Alliance says that Sweden’s new law “will further isolate sex workers, particularly migrants and trans people”. It will also be hard to enforce. Instead of expanding the Nordic model online, Sweden should be looking at better alternatives for both the real and digital worlds. Recently, in a church in Brussels, stalls offered lube, condoms and leaflets with titles such as: “So you’re dating a sex worker”; and “Anal health”. The event, arranged by the Belgian union of sex workers to mark International Whores’ Day, closed with a performance by a dominatrix, who stripped in front of the altar before hypnotising attendees with her whips. In 2022 Belgium became the first European country to fully decriminalise sex work (and the second globally after New Zealand). Last December it granted sex workers formal labour rights, entitling them to sick leave, maternity pay and pensions, and obliging brothels to get a permit and comply with health and safety standards. In the past few years, the Australian states of Victoria and Queensland have decriminalised sex work; South Africa and Thailand have drafted similar bills. In contrast to the Nordic model, this approach is supported by many academics, who argue that full decriminalisation helps to reduce stigma, deters police harassment and helps separate the willing from the coerced.
FOR DECADES Sweden was seen as the epitome of sexual freedom, so much so that President Dwight Eisenhower fulminated in 1960 that its people tended towards “sin, nudity, drunkenness”. In 1971 it followed Denmark to become the second country in the world to legalise all forms of pornography. Yet Sweden has been altogether more prudish when it comes to prostitution, having originated the so-called Nordic Model in 1999, which criminalised the purchase of sex, but not its sale, with the intention of reducing demand while protecting vulnerable women. This model has since spread widely. In the past decade, France, Ireland, Israel and the American state of Maine have all adopted it; Scotland is considering it. Now Sweden is trying to apply its real-world Nordic model to the digital world.On July 1st a new Swedish law will come into force that criminalises paying for live porn on sites like OnlyFans, the platform best known for its adult content, but not those who perform the online sex acts. Those breaking the new law face a penalty of up to a year in prison. The new law comes as sex workers increasingly embrace technology and move online. Many of those selling physical sex have already transitioned from soliciting on street corners to advertising on the internet. More recently the sale of virtual sex has gained momentum. Buying sex and sexual content has also become easier-and more accepted-than ever. Fully 14% of young Americans say they would consider selling content on OnlyFans; 16% say they would become a "sugar baby", selling companionship (and very often sex) to older men. In May 2024 one in ten British adults who were online visited Chaturbate, a live webcam-sex site. And the boundaries of what people consider to be sex work are becoming blurry, notes Teela Sanders of the University of Leicester. In Sweden, 8% of girls aged 15-19 say they have sent sexual content or met someone for sex in exchange for money often via the social-media app Snapchat. On the gay dating app Grindr, it is common to trade sex for gifts or drugs. In an age where sex and relationships are often more overtly transactional , many in Generation Z do not see what they are doing as "sex work" at all (unlike, for example, work in brothels, which is usually perceived as more exploitative). Only 56% of Britons aged 18-25 say that"sugaring" counts as sex work, compared with 70% of over-65s. These changes raise two important questions. Does the Swedish model work in the real world? And should it also apply to virtual sex work? Supporters of the Nordic model in relation to old-school prostitution argue that the purchase of sex is always exploitative, and that restrictions are needed not only to protect the vulnerable but also to crush demand. In this they hark back to abolitionist views of the 1920s, when the League of Nations sent undercover agents to infiltrate brothels, where they found examples of foreign women in debt bondage. Denying sex workers agency over their own bodies, they declared that many were "mentally abnormal" and of "poor heredity and poor environment" Though attitudes have softened since, some of the stigma and denial of agency persist in the Nordic model, which is based on the idea that many prostitutes go into sex work because of childhood abuse or poverty rather than free choice, and that they should be "rehabilitated". To be sure, there are many sex workers like Samuel Vahlund, who turned to prostitution after childhood bullying and a traumatic sexual assault destroyed his self-esteem, and because he needed money for drugs. At the time, he believed he was consenting, but would dissociate during sex with other men. “I hated myself,” he says. And similar trends appear to be the case for many of those entering online sex work. Meghan Donevan, a researcher at Talita, a Swedish charity, interviewed 120 people who had appeared in “commercial pornography”. Of those, 88% had been sexually abused as children and 69% had attempted suicide. Virtual sex work also creates the possibility for other sorts of harm. In thousands of webcam “studios” across Colombia, women work 12-hour shifts streaming from cramped, often filthy cubicles, keeping only a fraction of their earnings. “Doxxing”, whereby personal information is published on the internet, is common. “I know a lot of people who prefer old-school sex work because it’s safer,” says one Italian sex worker. Sweden’s government argues that since virtual sex work also carries the risk of harm, it should get the same legal treatment as the sale of actual sex. Moreover, says Nina Larson, Sweden’s minister for gender equality, “digital prostitution” can be a “gateway” to selling sex in person. Does the argument have merit? Supporters of the existing Nordic model tout a fall in the share of Swedish men who say they have ever paid for sex, from 14% in 1996 to 9% in 2017. They also point to falling street prostitution as proof of its success. But that may well be because sex work has moved underground or out of the country. The government reckons that 80% of Swedish men who pay for sex do so when they are abroad. Janna Davidson, Sweden’s national rapporteur on human trafficking and prostitution, argues that the most important effect of the Nordic model has been normative: Swedes are far more likely than their rich-world peers to say prostitution can “never be justified”. Yet there is little firm evidence that the Nordic model is achieving its main goals of protecting vulnerable sex workers or of putting a lasting dent in demand, especially since a growing proportion of men suspected of buying sex are under 30 Worse, the policies in force since 1999 may have distracted authorities from reducing the genuinely harmful aspects of the sex trade. This is because police often find it easier to arrest johns than to pursue pimps and traffickers. Prosecutions for buying sex have risen in Sweden, but there were no convictions for trafficking in 2024. Far from protecting vulnerable people, the existing laws may well be harming them. The stigmatisation of sex work means prostitutes and their clients are less likely to seek treatment for sexually transmitted infections. Migrants—who may make up 70-80% of those selling sex—are particularly vulnerable, because even though the sale of sex is not criminalised, it can be grounds for deportation. Paulina Bolton, who co-ordinates investigations into prostitution and human trafficking at the Swedish Gender Equality Agency, admits that support for victims is lacking. The European Sex Workers Rights Alliance says that Sweden’s new law “will further isolate sex workers, particularly migrants and trans people”. It will also be hard to enforce. Instead of expanding the Nordic model online, Sweden should be looking at better alternatives for both the real and digital worlds. Recently, in a church in Brussels, stalls offered lube, condoms and leaflets with titles such as: “So you’re dating a sex worker”; and “Anal health”. The event, arranged by the Belgian union of sex workers to mark International Whores’ Day, closed with a performance by a dominatrix, who stripped in front of the altar before hypnotising attendees with her whips. In 2022 Belgium became the first European country to fully decriminalise sex work (and the second globally after New Zealand). Last December it granted sex workers formal labour rights, entitling them to sick leave, maternity pay and pensions, and obliging brothels to get a permit and comply with health and safety standards. In the past few years, the Australian states of Victoria and Queensland have decriminalised sex work; South Africa and Thailand have drafted similar bills. In contrast to the Nordic model, this approach is supported by many academics, who argue that full decriminalisation helps to reduce stigma, deters police harassment and helps separate the willing from the coerced.
考点1:“criminalise the purchase of sex”应译为【将购买性服务行为入罪】 考点2:“live porn”推荐译为【在线色情直播】或【实时成人视频】 考点3:“sugar baby”推荐译为【甜心宝贝】 考点4:“doxxing”应译为【人肉搜索】或【恶意信息曝光】 考点5:“debt bondage”应译为【债役】或【债务奴役】 考点6:“abolitionist views”推荐译为【禁娼主义观点】或【废除主义立场】 考点7:“digital prostitution”应译为【数字性交易】或【数字性服务】 考点8:“sex trafficking”推荐译为【性交易人口贩运】 考点9:“International Whores’ Day”应统一译为【国际妓女节】 考点10:“dominatrix”推荐译为【女施虐者】或【女王调教者】 考点11:“decriminalise sex work”推荐译为【将性工作非刑事化】或【废除刑事处罚】 考点12:“police harassment”推荐译为【警方骚扰】或【执法骚扰】 考点13:“rehabilitated”推荐译为表述【改造】【改过自新】、【重新融入社会】等,不可直译为【康复】 考点14: “Only 56% of Britons aged 18-25 say that"sugaring" counts as sex work, compared with 70% of over-65s.”中“sugaring”推荐译为【包养】,不应译为【糖交】 考点15:“poor heredity and poor environment”推荐翻译为【出身和生长环境很差】
764b8
新闻资讯
观点评论
74
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 在当代智能交通技术快速演进的背景下,端到端自动驾驶系统基于深度神经网络的方案日益成为研究与工业界的焦点。相较于传统架构中将感知、决策制定与运动控制明确划分为多个独立的功能模块,这种端到端方法力图将整条自动驾驶处理链整合进一个统一的神经网络模型中,直接由原始多模态传感器输入(如摄像头图像、雷达点云或车辆历史状态)输出连续的驾驶指令(如转向角、加减速命令、急刹车信号),从而大幅减少中间接口的工程复杂性,简化调试流程,并显著降低感知与控制之间可能发生的不一致性所带来的系统隐患。 然而,端到端方法的实际部署远非“结构简化”这么简单。要在工业级别的自动驾驶平台上投入应用,其首先需要满足一个关键前提:实时性。也就是说,系统从输入图像接收到生成控制命令的延迟必须足够低,以满足高速行驶车辆的动态响应要求。在满足这一严苛条件的过程中,研究者不仅要在算法层面对神经网络进行结构压缩与推理加速,如使用稀疏注意力机制、通道剪枝、深度可分离卷积或层级特征重用等手段,同时还需从底层软硬件协同角度入手,针对不同部署平台(如数据中心、车载边缘设备)设计异构优化策略。例如,在数据中心的训练和推理阶段,研究团队普遍采用TensorRT或ONNX Runtime等主流高效推理引擎,配合NVIDIA GPU的Tensor Core或AMD ROCm平台,进行混合精度训练与INT8低延迟推理;而对于部署在车载终端的轻量级边缘模型,则常借助FPGA的高度可编程性和专用NPU(如寒武纪、地平线)的流式处理特性,通过深度流水线调度、内存片上缓存重用等策略将感知延迟压缩至毫秒级别甚至更低。 在网络结构方面,近年来兴起的视觉–语言–动作(VLA)三模态融合架构,为端到端自动驾驶系统带来了前所未有的语义泛化能力。其核心在于构建一个支持图像感知、语言理解与动作决策统一表示的多模态融合网络。该结构通常包含以下几个关键子模块: (1)视觉编码器模块,基于改进后的ResNet-101、EfficientNetV2或新兴的Vision Transformer(ViT/DeiT),负责从原始RGB图像中提取尺度丰富、时空相关的图像特征张量; (2)语言编码器模块,通常采用基于Transformer或GPT结构的序列建模器,将高层次的语义指令(如“在下一个交通灯左转后并道”或“避让前方施工车辆”)转化为可嵌入的高维语义向量; (3)多模态对齐与融合模块,该模块常采用Cross-Attention机制对视觉与语言特征进行相互对齐,随后通过多个堆叠的Transformer Encoder层提取跨模态语义表征; (4)动作解码器,使用如一维卷积网络(TemporalConvNet)、LSTM 或 GRU 等时序建模单元预测未来两秒内每100毫秒一个离散采样点的轨迹控制命令。 值得注意的是,这类多模态架构中语言的加入使得驾驶系统具备某种程度的“高层可控性”,即用户或系统设计者可以通过语言指令引导行为意图。例如,在城市工况下可以通过指令指定特殊策略(如“靠右行驶”、“优先保持直行”等),显著提高驾驶系统的可解释性与策略灵活性。 数据是决定模型性能的根本性因素。一个高质量、具有代表性、可扩展的数据集,不仅关系到模型训练阶段的收敛效果,更对最终泛化性能起到决定性作用。当前的数据集建设方法通常包括两大类:仿真合成数据与真实车辆采集数据。前者依赖高保真的模拟平台(如CARLA、LGSVL、MetaDrive等),可生成多样化的极端天气(暴雪、浓雾、夜间雨天)、稀有交通行为(逆行、鬼探头、非规范掉头)与异常路况(车道消失、交通灯失效等),且每一帧数据都自带准确的语义分割、检测框、轨迹注释等标签;后者则往往需要动用多车队列采集,借助高精度IMU/GPS组合导航系统、128线激光雷达、多角度鱼眼摄像头等多传感器融合技术,获得同步、高分辨率的传感信息与精确标注。为了保证训练数据与验证数据分布的一致性,标注团队通常需要构建细致的标注手册,并定期进行交叉校验,以降低“标签漂移”或“语义漂移”的风险。 在数据预处理过程中,还常引入诸如几何变换(旋转、缩放、仿射扰动)、光照变化模拟、色彩扰动、镜像反转、目标遮挡等增强策略,以提升模型在不同场景与摄像机角度下的鲁棒性与泛化能力。 模型训练阶段一般分为两个主要阶段:第一阶段为有监督的预训练,采用大规模公开数据集(如nuScenes、Waymo Open Dataset、Argoverse 2.0等)进行目标识别、轨迹预测或行为分类任务的学习;第二阶段则引入强化学习(RL)进行策略微调,通过构建闭环仿真环境、定义奖励函数,并使用如DDPG、SAC、PPO等策略梯度算法训练Agent,在具体交通情景下学习最优行为边界。这一阶段的目标并非单纯提高轨迹准确率,而是让系统在高维状态空间中掌握对复杂交通互动关系的动态决策能力,如在车流密集交汇区域如何插队、在无信号交叉路口如何让行等。 模型部署阶段则是将训练完成的系统集成至车载平台中的关键一环。首先,需要在嵌入式系统上安装适配的实时操作系统(如QNX、Linux RT Patch)与自研中间件,负责调度处理器资源、协调各传感器采样频率,保证雷达、摄像头、超声波等设备在统一时间线上输出。系统需持续监控推理帧率、内存占用、模型漂移率、控制响应延迟等鲁棒性指标;一旦检测到关键性能指标跌出安全阈值区间,便会触发“安全降级”机制,即自动将控制权从主系统转交至冗余控制器或传统的规则系统(如基于PID的紧急停车控制),以避免“算法在正常情况下如火箭般展现性能,而在边缘情况下一败涂地”的风险。 此外,在真实道路环境中进行测试时,还必须构建完善的回放与重构系统。每辆测试车需具备本地日志记录与高速无线数据上传能力,所有运行日志需回传至云端平台,由仿真回放系统(如RESim或OpenReplay)进行逐帧复刻与行为审计。这类系统常用于挖掘“边缘案例”——即发生概率极低但具有极大安全风险的特殊情况,如小孩突然冲入路口、车辆遭遇紧急制动但被后车追尾等。 最终,在正式版本上线前,工程团队通常需进行大规模在线A/B测试,以对比不同版本模型在用户体验、安全冗余、策略鲁棒性等方面的实际差异。只有当所有指标显著优于旧版本且无新增失败案例时,系统才会完成灰度上线,确保每一次更新都稳定、透明、可追溯,从而实现真正可落地的“全栈端到端自动驾驶体验”。
在当代智能交通技术快速演进的背景下,端到端自动驾驶系统基于深度神经网络的方案日益成为研究与工业界的焦点。相较于传统架构中将感知、决策制定与运动控制明确划分为多个独立的功能模块,这种端到端方法力图将整条自动驾驶处理链整合进一个统一的神经网络模型中,直接由原始多模态传感器输入(如摄像头图像、雷达点云或车辆历史状态)输出连续的驾驶指令(如转向角、加减速命令、急刹车信号),从而大幅减少中间接口的工程复杂性,简化调试流程,并显著降低感知与控制之间可能发生的不一致性所带来的系统隐患。 然而,端到端方法的实际部署远非“结构简化”这么简单。要在工业级别的自动驾驶平台上投入应用,其首先需要满足一个关键前提:实时性。也就是说,系统从输入图像接收到生成控制命令的延迟必须足够低,以满足高速行驶车辆的动态响应要求。在满足这一严苛条件的过程中,研究者不仅要在算法层面对神经网络进行结构压缩与推理加速,如使用稀疏注意力机制、通道剪枝、深度可分离卷积或层级特征重用等手段,同时还需从底层软硬件协同角度入手,针对不同部署平台(如数据中心、车载边缘设备)设计异构优化策略。例如,在数据中心的训练和推理阶段,研究团队普遍采用TensorRT或ONNX Runtime等主流高效推理引擎,配合NVIDIA GPU的Tensor Core或AMD ROCm平台,进行混合精度训练与INT8低延迟推理;而对于部署在车载终端的轻量级边缘模型,则常借助FPGA的高度可编程性和专用NPU(如寒武纪、地平线)的流式处理特性,通过深度流水线调度、内存片上缓存重用等策略将感知延迟压缩至毫秒级别甚至更低。 在网络结构方面,近年来兴起的视觉–语言–动作(VLA)三模态融合架构,为端到端自动驾驶系统带来了前所未有的语义泛化能力。其核心在于构建一个支持图像感知、语言理解与动作决策统一表示的多模态融合网络。该结构通常包含以下几个关键子模块: (1)视觉编码器模块,基于改进后的ResNet-101、EfficientNetV2或新兴的Vision Transformer(ViT/DeiT),负责从原始RGB图像中提取尺度丰富、时空相关的图像特征张量; (2)语言编码器模块,通常采用基于Transformer或GPT结构的序列建模器,将高层次的语义指令(如“在下一个交通灯左转后并道”或“避让前方施工车辆”)转化为可嵌入的高维语义向量; (3)多模态对齐与融合模块,该模块常采用Cross-Attention机制对视觉与语言特征进行相互对齐,随后通过多个堆叠的Transformer Encoder层提取跨模态语义表征; (4)动作解码器,使用如一维卷积网络(TemporalConvNet)、LSTM 或 GRU 等时序建模单元预测未来两秒内每100毫秒一个离散采样点的轨迹控制命令。 值得注意的是,这类多模态架构中语言的加入使得驾驶系统具备某种程度的“高层可控性”,即用户或系统设计者可以通过语言指令引导行为意图。例如,在城市工况下可以通过指令指定特殊策略(如“靠右行驶”、“优先保持直行”等),显著提高驾驶系统的可解释性与策略灵活性。 数据是决定模型性能的根本性因素。一个高质量、具有代表性、可扩展的数据集,不仅关系到模型训练阶段的收敛效果,更对最终泛化性能起到决定性作用。当前的数据集建设方法通常包括两大类:仿真合成数据与真实车辆采集数据。前者依赖高保真的模拟平台(如CARLA、LGSVL、MetaDrive等),可生成多样化的极端天气(暴雪、浓雾、夜间雨天)、稀有交通行为(逆行、鬼探头、非规范掉头)与异常路况(车道消失、交通灯失效等),且每一帧数据都自带准确的语义分割、检测框、轨迹注释等标签;后者则往往需要动用多车队列采集,借助高精度IMU/GPS组合导航系统、128线激光雷达、多角度鱼眼摄像头等多传感器融合技术,获得同步、高分辨率的传感信息与精确标注。为了保证训练数据与验证数据分布的一致性,标注团队通常需要构建细致的标注手册,并定期进行交叉校验,以降低“标签漂移”或“语义漂移”的风险。 在数据预处理过程中,还常引入诸如几何变换(旋转、缩放、仿射扰动)、光照变化模拟、色彩扰动、镜像反转、目标遮挡等增强策略,以提升模型在不同场景与摄像机角度下的鲁棒性与泛化能力。 模型训练阶段一般分为两个主要阶段:第一阶段为有监督的预训练,采用大规模公开数据集(如nuScenes、Waymo Open Dataset、Argoverse 2.0等)进行目标识别、轨迹预测或行为分类任务的学习;第二阶段则引入强化学习(RL)进行策略微调,通过构建闭环仿真环境、定义奖励函数,并使用如DDPG、SAC、PPO等策略梯度算法训练Agent,在具体交通情景下学习最优行为边界。这一阶段的目标并非单纯提高轨迹准确率,而是让系统在高维状态空间中掌握对复杂交通互动关系的动态决策能力,如在车流密集交汇区域如何插队、在无信号交叉路口如何让行等。 模型部署阶段则是将训练完成的系统集成至车载平台中的关键一环。首先,需要在嵌入式系统上安装适配的实时操作系统(如QNX、Linux RT Patch)与自研中间件,负责调度处理器资源、协调各传感器采样频率,保证雷达、摄像头、超声波等设备在统一时间线上输出。系统需持续监控推理帧率、内存占用、模型漂移率、控制响应延迟等鲁棒性指标;一旦检测到关键性能指标跌出安全阈值区间,便会触发“安全降级”机制,即自动将控制权从主系统转交至冗余控制器或传统的规则系统(如基于PID的紧急停车控制),以避免“算法在正常情况下如火箭般展现性能,而在边缘情况下一败涂地”的风险。 此外,在真实道路环境中进行测试时,还必须构建完善的回放与重构系统。每辆测试车需具备本地日志记录与高速无线数据上传能力,所有运行日志需回传至云端平台,由仿真回放系统(如RESim或OpenReplay)进行逐帧复刻与行为审计。这类系统常用于挖掘“边缘案例”——即发生概率极低但具有极大安全风险的特殊情况,如小孩突然冲入路口、车辆遭遇紧急制动但被后车追尾等。 最终,在正式版本上线前,工程团队通常需进行大规模在线A/B测试,以对比不同版本模型在用户体验、安全冗余、策略鲁棒性等方面的实际差异。只有当所有指标显著优于旧版本且无新增失败案例时,系统才会完成灰度上线,确保每一次更新都稳定、透明、可追溯,从而实现真正可落地的“全栈端到端自动驾驶体验”。
考点1:“深度可分离卷积 ”必须译为Depthwise Separable Convolution,固定译法 考点2:“灰度上线”应该译为gray release或者dark launch,行业内熟知的说法
77596
学术论文
应用学科
196
翻译,英文翻译成中文。不要输出译文以外的内容。 以下是你本次的任务: Confidentiality 1. Each Party undertakes with the other Party that it shall treat as strictly confidential all information received or obtained by it or its employees agents or advisers as a result of ente-ring into or performing this Agreement including information relating to the provisions of this Agreement, the negotiations leading up to this Agreement, the subject matter of this Agreement or the business or affairs of the other Party or any member of the other Party's group of companies and that it shall not at any time hereafter make use of or disclose or divulge to any person any such information andshall use its best endeavors to prevent the publication or disclosure of any such information. 2. From and after the date of this Agreement, Party A shall, and shall cause its affiliates and successors to, use the same efforts to maintain the confidentiality of any Confidential Information as Party A used to maintain the confidentiality of such information prior to the date hereof.Notwithstanding the foregoing, Party A and its affiliates and successors shall after prior notice to, and consultation with Party B, be permitted to disclose any such Confidential Information to the extent legally required or necessary for obtaining appropriate regulatory licenses or approvals.3. All Proprietary Information disclosed by either Party or its affiliates to the CJV in accordance with the provisions of this Contract and the Technology License Agreement and/ or the Trademark License Agreement to be entered into between an affiliate of Party B and the CJV on or about the date hereof substantially in the form attached hereto as Appendix X shall be used by the CJV and its personnel solely for the CIV's account and purposes. Each Party and any of its affiliates shall maintain the secrecy of all Proprietary.Information that may be disclosed or furnished to it by the CJV or the other Party and its affiliates, and neither of the Parties or their affiliates shall disclose or reveal any such Proprietary Information to any third party without explicit written authorization from the other Party. Any Proprietary.Information obtained by the CIV or a Party or its affiliates may be disclosed only to the designated employees of the CIV or that Party and its affiliates whose duties so require for the implementation of this Contract. The CJV and each Party and its affiliates shall take all reasonable precautions(including the conclusion of confidentiality contracts with each such employee)to prevent such employees from using and disclosing the Proprietary Information in contravention of this Article 3. 4."Confidential Information"means all documents.software and documentation, reports, financial or other data, records, forms, tools.products, services, methodologies, present and future research, technical knowledge, marketing plans, trade secrets, and other materials obtained by Consultant and Client from each other in the course of performing any Services, whether tangible or intangible and whether or not stored, compiled or memorized physically, electronically, graphically, in writing, or by any means now known or later invented. Confidential Information includes without limitation records and information and Consultant's Information (i)that has been marked as proprietary or confidential; (ii) whose confidential nature has been made known by Client or Consultant; or (iii) that due to its character and nature, a reasonable person under like circumstances would treat as confidential. Notwithstanding the foregoing. Confidential Information does not include information which (i) is already known to the recipient at the time of disclosure, (ii) is or becomes publicly known through no wrongful act or failure of the recipient; (iii) is independently developed by the recipient without benefit of the other party's Confidential Information.or (iv) is received from a third party which is not under and does not thereby breach an obligation of confidentiality. Each party agrees to protect the other's Confidential Information at all times and in the same manner as each protects the confidentiality of its own proprietary and confidential materials.but in no event with less than a reasonable standard of care. Consultant will deliver to Client all Confidential Information and all copies thereof (and all other property obtained from or through Client) when Client requests the same or immediately upon termination of this Agreement, whichever occur searlier, except for one copy thereof that Consultant, may retain for its records. Neither party shall, except with respect to those of its employees with a need to know under this Agreement, use or disclose to any person firm or entity any Confidential Information of the other party without such other party's express, prior written permission; provided, however, that not withstanding the foregoing, Consultant may disclose Confidential Information to the extent that it is required to be disclosed pursuant to a statutory or regulatory provision or court order. The confidentiality restrictions and obligations imposed by this Section 4 shall terminate two(2)years after the expiration or termination of this Agreement. Definition and Interpretation 1. "Affiliate" means any person or company that directly or indirectly controls a Party or is directly or indirectly controlled by a Party, including a Party's parent or subsidiary, or is under direct or indirect common control with such Party. For the purpose of this Agreement, "control" shall mean either the ownership of fifty per cent (50%) or more of the ordinary share capital of the company carrying the right to vote at general meetings or the power to nominate a majority of the board of directors of the Company."Proprietary Know-how" shall mean processes,methods and manufacturing techniques, experience and other information and materials including but not limited to the Technical Information and Technical Assistance supplied or rendered by the Licensor to the Licensee here under which have been developed by and are known to the Licensor on the date here of and/or which may be further developed by the Licensor or become known to it during the continuance of this Agreement excepting, howeve rany secret know-how acquired by the Licensor from third parties which the Licensor is precluded from disclosing to the Licensee. 3. ""Proprietary Information" means the information, whether patentable or not, disclosed to the CJV by either Party or its Affiliates or disclosed by the CJV to either Party or its Affiliates during the term of this Contract,including technology, inventions, creations, know-how, formulations recipes, specifications, designs, methods, processes, techniques, data, rights devices, drawings, instructions, expertise, trade practices, trade secrets and such commercial, economic, financial or other information as is generally treated as confidential by the disclosing Party, its Affiliates, or the CJV, as the case may be; provided that when such information is in unwritten or intangible form, the disclosing Party, its Affiliates or the CJV shall, within one month of making the disclosure, provide the other Party and/or the CJV with a written confirmation that such information constitutes its Proprietary Information. 4. "Encumbrances" include any option, right to acquire, right of preemption mortgage, charge, pledge, lien, hypothecation, title creation, right of set-off counter claim, trust arrangement or other security or any equity or restriction(including any relevant restriction imposed under the relevant law). 5. In this Agreement, unless the context otherwise requires: a) headings are for convenience only and shall not affect the interpretation of this Agreement; b) words importing the singular include the plural and vice versa;c) words importing a gender include any gender;d) an expression importing a natural person includes any company partnership, joint venture, association, corporation or other body corporate and any governmental agency; e)a reference to any law, regulation or rule includes all laws, regulations, or rules amending, consolidating or replacing them, and a reference to a law includes all regulations and rules under that law; f) a reference to a document includes an amendment or supplement to, or replacement or novation of, that document;g)a reference to a party to any document includes that party's successors and permitted assigns; h) a reference to an agreement includes an undertaking, agreement or legally enforceable arrangement or understanding whether or not in writing,i) a warranty, representation, undertaking, indemnity, covenant or agreement on the part of two or more persons binds them jointly and severally; and j) the schedules, annexures and appendices to this Agreement shall form an integral part of this Agreement. 1. "Control" means the power of any entity to direct the affairs of another entity whether by any of the following means or otherwise:.a) beneficial ownership of or entitlement to acquire shares or assets of the other; b) power to exercise voting rights in relation to the other;c)power to appoint members of the supervisory board, board of directors or bodies legally representing the other; or d) power to operate, direct or manage the general affairs of the other.2. "Technical Information" shall mean all technical data, drawings, designs,formulae, specifications, processes, methods of manufacture, computer programs or other software or technical documents and similar intellectual property rights of or developed by the Licensor as of the date hereof or to be developed by the Licensor during the term of this Agreement relating to the Product, including improvements thereto. 3. The headings used in this Contract are for ease of reference only, and in no event shall the substance of any paragraph or the intent of the Parties be interpreted or controlled by such headings.4.References to: a)Clauses and Schedules are to clauses in and schedules to this Agreement(unless the context otherwise requires) and the Recitals and Schedules to this Agreement shall be deemed to form part of this Agreement; b)"person" shall include body corporate》 unincorporated association and partnership (whether or not having separate legal personality);c) writing shall include any methods of producing or reproducing words in a legible and non-transitory form; and d) masculine gender shall include the feminine and neuter and the singular number shall include the plural and vice versa.
Confidentiality 1. Each Party undertakes with the other Party that it shall treat as strictly confidential all information received or obtained by it or its employees agents or advisers as a result of ente-ring into or performing this Agreement including information relating to the provisions of this Agreement, the negotiations leading up to this Agreement, the subject matter of this Agreement or the business or affairs of the other Party or any member of the other Party's group of companies and that it shall not at any time hereafter make use of or disclose or divulge to any person any such information andshall use its best endeavors to prevent the publication or disclosure of any such information. 2. From and after the date of this Agreement, Party A shall, and shall cause its affiliates and successors to, use the same efforts to maintain the confidentiality of any Confidential Information as Party A used to maintain the confidentiality of such information prior to the date hereof.Notwithstanding the foregoing, Party A and its affiliates and successors shall after prior notice to, and consultation with Party B, be permitted to disclose any such Confidential Information to the extent legally required or necessary for obtaining appropriate regulatory licenses or approvals.3. All Proprietary Information disclosed by either Party or its affiliates to the CJV in accordance with the provisions of this Contract and the Technology License Agreement and/ or the Trademark License Agreement to be entered into between an affiliate of Party B and the CJV on or about the date hereof substantially in the form attached hereto as Appendix X shall be used by the CJV and its personnel solely for the CIV's account and purposes. Each Party and any of its affiliates shall maintain the secrecy of all Proprietary.Information that may be disclosed or furnished to it by the CJV or the other Party and its affiliates, and neither of the Parties or their affiliates shall disclose or reveal any such Proprietary Information to any third party without explicit written authorization from the other Party. Any Proprietary.Information obtained by the CIV or a Party or its affiliates may be disclosed only to the designated employees of the CIV or that Party and its affiliates whose duties so require for the implementation of this Contract. The CJV and each Party and its affiliates shall take all reasonable precautions(including the conclusion of confidentiality contracts with each such employee)to prevent such employees from using and disclosing the Proprietary Information in contravention of this Article 3. 4."Confidential Information"means all documents.software and documentation, reports, financial or other data, records, forms, tools.products, services, methodologies, present and future research, technical knowledge, marketing plans, trade secrets, and other materials obtained by Consultant and Client from each other in the course of performing any Services, whether tangible or intangible and whether or not stored, compiled or memorized physically, electronically, graphically, in writing, or by any means now known or later invented. Confidential Information includes without limitation records and information and Consultant's Information (i)that has been marked as proprietary or confidential; (ii) whose confidential nature has been made known by Client or Consultant; or (iii) that due to its character and nature, a reasonable person under like circumstances would treat as confidential. Notwithstanding the foregoing. Confidential Information does not include information which (i) is already known to the recipient at the time of disclosure, (ii) is or becomes publicly known through no wrongful act or failure of the recipient; (iii) is independently developed by the recipient without benefit of the other party's Confidential Information.or (iv) is received from a third party which is not under and does not thereby breach an obligation of confidentiality. Each party agrees to protect the other's Confidential Information at all times and in the same manner as each protects the confidentiality of its own proprietary and confidential materials.but in no event with less than a reasonable standard of care. Consultant will deliver to Client all Confidential Information and all copies thereof (and all other property obtained from or through Client) when Client requests the same or immediately upon termination of this Agreement, whichever occur searlier, except for one copy thereof that Consultant, may retain for its records. Neither party shall, except with respect to those of its employees with a need to know under this Agreement, use or disclose to any person firm or entity any Confidential Information of the other party without such other party's express, prior written permission; provided, however, that not withstanding the foregoing, Consultant may disclose Confidential Information to the extent that it is required to be disclosed pursuant to a statutory or regulatory provision or court order. The confidentiality restrictions and obligations imposed by this Section 4 shall terminate two(2)years after the expiration or termination of this Agreement. Definition and Interpretation 1. "Affiliate" means any person or company that directly or indirectly controls a Party or is directly or indirectly controlled by a Party, including a Party's parent or subsidiary, or is under direct or indirect common control with such Party. For the purpose of this Agreement, "control" shall mean either the ownership of fifty per cent (50%) or more of the ordinary share capital of the company carrying the right to vote at general meetings or the power to nominate a majority of the board of directors of the Company."Proprietary Know-how" shall mean processes,methods and manufacturing techniques, experience and other information and materials including but not limited to the Technical Information and Technical Assistance supplied or rendered by the Licensor to the Licensee here under which have been developed by and are known to the Licensor on the date here of and/or which may be further developed by the Licensor or become known to it during the continuance of this Agreement excepting, howeve rany secret know-how acquired by the Licensor from third parties which the Licensor is precluded from disclosing to the Licensee. 3. ""Proprietary Information" means the information, whether patentable or not, disclosed to the CJV by either Party or its Affiliates or disclosed by the CJV to either Party or its Affiliates during the term of this Contract,including technology, inventions, creations, know-how, formulations recipes, specifications, designs, methods, processes, techniques, data, rights devices, drawings, instructions, expertise, trade practices, trade secrets and such commercial, economic, financial or other information as is generally treated as confidential by the disclosing Party, its Affiliates, or the CJV, as the case may be; provided that when such information is in unwritten or intangible form, the disclosing Party, its Affiliates or the CJV shall, within one month of making the disclosure, provide the other Party and/or the CJV with a written confirmation that such information constitutes its Proprietary Information. 4. "Encumbrances" include any option, right to acquire, right of preemption mortgage, charge, pledge, lien, hypothecation, title creation, right of set-off counter claim, trust arrangement or other security or any equity or restriction(including any relevant restriction imposed under the relevant law). 5. In this Agreement, unless the context otherwise requires: a) headings are for convenience only and shall not affect the interpretation of this Agreement; b) words importing the singular include the plural and vice versa;c) words importing a gender include any gender;d) an expression importing a natural person includes any company partnership, joint venture, association, corporation or other body corporate and any governmental agency; e)a reference to any law, regulation or rule includes all laws, regulations, or rules amending, consolidating or replacing them, and a reference to a law includes all regulations and rules under that law; f) a reference to a document includes an amendment or supplement to, or replacement or novation of, that document;g)a reference to a party to any document includes that party's successors and permitted assigns; h) a reference to an agreement includes an undertaking, agreement or legally enforceable arrangement or understanding whether or not in writing,i) a warranty, representation, undertaking, indemnity, covenant or agreement on the part of two or more persons binds them jointly and severally; and j) the schedules, annexures and appendices to this Agreement shall form an integral part of this Agreement. 1. "Control" means the power of any entity to direct the affairs of another entity whether by any of the following means or otherwise:.a) beneficial ownership of or entitlement to acquire shares or assets of the other; b) power to exercise voting rights in relation to the other;c)power to appoint members of the supervisory board, board of directors or bodies legally representing the other; or d) power to operate, direct or manage the general affairs of the other.2. "Technical Information" shall mean all technical data, drawings, designs,formulae, specifications, processes, methods of manufacture, computer programs or other software or technical documents and similar intellectual property rights of or developed by the Licensor as of the date hereof or to be developed by the Licensor during the term of this Agreement relating to the Product, including improvements thereto. 3. The headings used in this Contract are for ease of reference only, and in no event shall the substance of any paragraph or the intent of the Parties be interpreted or controlled by such headings.4.References to: a)Clauses and Schedules are to clauses in and schedules to this Agreement(unless the context otherwise requires) and the Recitals and Schedules to this Agreement shall be deemed to form part of this Agreement; b)"person" shall include body corporate》 unincorporated association and partnership (whether or not having separate legal personality);c) writing shall include any methods of producing or reproducing words in a legible and non-transitory form; and d) masculine gender shall include the feminine and neuter and the singular number shall include the plural and vice versa.
考点1:“use its best endeavors”建议译为“竭力。 考点2:“without explicit written authorization”需译为“未经明示书面授权”。 考点3:“affiliates and successors”推荐译为“关联公司和承继人”。 考点4:“solely for the... purposes”译为“仅供……之目的”。 考点5:“reasonable precautions”建议译为“合理的预防措施”。 考点6:“terminate two (2) years after”译为“于……后二(2)年终止”。
78481
垂类场景
法律
120
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 三维(3D)复合材料是一种基于三维机织整体成型技术的先进材料,其结构由上下层织物面层与交叉呈"8"字形的绒经纤维组成。这种独特的设计通过Z向纤维在毛细作用下形成三维整体框架结构,利用织物中空结构瞬间吸收树脂,赋予材料轻质高强、抗分层剥离的优异特性。材料的制造工艺是采用了环氧树脂注入增强筋孔并与三维织物复合模压固化,成品面密度为4kg/m²,可承载500kg/m²荷载及45m/s风压。通过调节织物密度、层数及空间形态,可以灵活设计适配不同强度需求,同时中空层可通过填充泡沫或预埋功能组件实现隔音、阻燃等附加性能。 在应用方面,三维(3D)复合材料凭借其卓越的强度、增强的刚度、延长的抗疲劳性和低密度等优势,在航空航天、轨道交通和生物医学领域发挥着重要作用。然而,复合材料的疲劳损伤行为具有高度复杂性和不确定性,其潜在损伤机制至今未能充分阐明。但研究表明,在拉伸-压缩疲劳载荷下的疲劳寿命明显短于拉伸-拉伸疲劳,其损伤过程通常始于基体开裂,继而发展为分层、微屈曲、纤维扭结,最终导致纤维断裂。基于前人的研究,复合材料的疲劳性能与传统金属材料有很大程度上的不同,失效机理和失效模式也比金属材料复杂得多。这种复杂的失效机理与传统金属材料有本质区别,金属材料通常以单一主控裂纹为主导,而复合材料则可能出现多种损伤模式同时存在的情况。 目前,研究三维复合材料疲劳性能的主要方法是实验研究,但这种方法耗时费力,且难以全面评估不同应力水平下的疲劳寿命。为克服这些限制,将现有的复合材料疲劳模型分类,主要包括三类:疲劳寿命模型、预测残余刚度/强度的现象学模型和渐进损伤模型。疲劳寿命模型主要基于应力水平-疲劳循环关系表征复合材料的疲劳性能,该方法需要大量的实验数据,没有关注复合材料结构在疲劳过程中的实际损伤机理,它只能通过预测特定类型复合材料结构在特定工况下的疲劳寿命。而预测残余刚度/强度的现象学模型基于复合材料的疲劳理论,没有关注疲劳过程中的实际损伤机理,缺乏对纤维断裂、基体开裂、分层等多种损伤模式的全面描述。渐进式损伤模型应用疲劳失效准则来确定和定量描述失效模式,例如基体开裂、基体压缩、纤维断裂、纤维压缩、基体/纤维剪切和分层。同时,从疲劳性能出发,建立了各主要方向疲劳载荷的残余强度和残余刚度退化模型,用于材料性能退化分析,为有限元模型中的刚度矩阵退化提供了依据。该预测模型可以描述疲劳加载过程中的具体损伤模式,有利于分析损伤机理。此外,基于复合材料疲劳损伤理论,适用于预测不同疲劳载荷条件和层形的寿命。 随着计算机技术的进步,有限元分析(FEA)在复合材料研究中的应用日益广泛。有限元分析的基本思想是将复杂问题离散化为有限个互连的子域进行分析求解。在复合材料领域,研究者已成功将Hashin准则与Murakami-Ohno理论相结合,建立了渐进损伤分析模型,准确地预测了三维编织复合材料的单轴拉伸特性。此外,有限元分析还被应用于热物理性能研究,比较不同结构复合材料的导热系数,并分析各参数对导热性能的影响规律。值得注意的是,随着有限元分析技术的发展,该方法已从早期的拉伸和热物理性能表征,进而扩展到复合材料疲劳行为的研究领域。例如,通过建立代表性体积单元(RVE)预测模型,研究者成功模拟了二维编织复合材料在双轴交替循环载荷下的疲劳寿命。另外,还有研究者基于介尺度代表性晶胞(RUC)模型分析了三维四向编织复合材料的疲劳特性,并研究了编织参数对线性或简化应力状态下疲劳行为的影响。 这些研究进展为三维复合材料的性能优化和工程应用提供了重要理论基础,特别是通过计算机模拟技术,可以在设计阶段预测材料性能,有效避免过度设计和安全系数过大等问题,对推动三维(3D)复合材料在各工业领域的应用具有重要意义。
三维(3D)复合材料是一种基于三维机织整体成型技术的先进材料,其结构由上下层织物面层与交叉呈"8"字形的绒经纤维组成。这种独特的设计通过Z向纤维在毛细作用下形成三维整体框架结构,利用织物中空结构瞬间吸收树脂,赋予材料轻质高强、抗分层剥离的优异特性。材料的制造工艺是采用了环氧树脂注入增强筋孔并与三维织物复合模压固化,成品面密度为4kg/m²,可承载500kg/m²荷载及45m/s风压。通过调节织物密度、层数及空间形态,可以灵活设计适配不同强度需求,同时中空层可通过填充泡沫或预埋功能组件实现隔音、阻燃等附加性能。 在应用方面,三维(3D)复合材料凭借其卓越的强度、增强的刚度、延长的抗疲劳性和低密度等优势,在航空航天、轨道交通和生物医学领域发挥着重要作用。然而,复合材料的疲劳损伤行为具有高度复杂性和不确定性,其潜在损伤机制至今未能充分阐明。但研究表明,在拉伸-压缩疲劳载荷下的疲劳寿命明显短于拉伸-拉伸疲劳,其损伤过程通常始于基体开裂,继而发展为分层、微屈曲、纤维扭结,最终导致纤维断裂。基于前人的研究,复合材料的疲劳性能与传统金属材料有很大程度上的不同,失效机理和失效模式也比金属材料复杂得多。这种复杂的失效机理与传统金属材料有本质区别,金属材料通常以单一主控裂纹为主导,而复合材料则可能出现多种损伤模式同时存在的情况。 目前,研究三维复合材料疲劳性能的主要方法是实验研究,但这种方法耗时费力,且难以全面评估不同应力水平下的疲劳寿命。为克服这些限制,将现有的复合材料疲劳模型分类,主要包括三类:疲劳寿命模型、预测残余刚度/强度的现象学模型和渐进损伤模型。疲劳寿命模型主要基于应力水平-疲劳循环关系表征复合材料的疲劳性能,该方法需要大量的实验数据,没有关注复合材料结构在疲劳过程中的实际损伤机理,它只能通过预测特定类型复合材料结构在特定工况下的疲劳寿命。而预测残余刚度/强度的现象学模型基于复合材料的疲劳理论,没有关注疲劳过程中的实际损伤机理,缺乏对纤维断裂、基体开裂、分层等多种损伤模式的全面描述。渐进式损伤模型应用疲劳失效准则来确定和定量描述失效模式,例如基体开裂、基体压缩、纤维断裂、纤维压缩、基体/纤维剪切和分层。同时,从疲劳性能出发,建立了各主要方向疲劳载荷的残余强度和残余刚度退化模型,用于材料性能退化分析,为有限元模型中的刚度矩阵退化提供了依据。该预测模型可以描述疲劳加载过程中的具体损伤模式,有利于分析损伤机理。此外,基于复合材料疲劳损伤理论,适用于预测不同疲劳载荷条件和层形的寿命。 随着计算机技术的进步,有限元分析(FEA)在复合材料研究中的应用日益广泛。有限元分析的基本思想是将复杂问题离散化为有限个互连的子域进行分析求解。在复合材料领域,研究者已成功将Hashin准则与Murakami-Ohno理论相结合,建立了渐进损伤分析模型,准确地预测了三维编织复合材料的单轴拉伸特性。此外,有限元分析还被应用于热物理性能研究,比较不同结构复合材料的导热系数,并分析各参数对导热性能的影响规律。值得注意的是,随着有限元分析技术的发展,该方法已从早期的拉伸和热物理性能表征,进而扩展到复合材料疲劳行为的研究领域。例如,通过建立代表性体积单元(RVE)预测模型,研究者成功模拟了二维编织复合材料在双轴交替循环载荷下的疲劳寿命。另外,还有研究者基于介尺度代表性晶胞(RUC)模型分析了三维四向编织复合材料的疲劳特性,并研究了编织参数对线性或简化应力状态下疲劳行为的影响。 这些研究进展为三维复合材料的性能优化和工程应用提供了重要理论基础,特别是通过计算机模拟技术,可以在设计阶段预测材料性能,有效避免过度设计和安全系数过大等问题,对推动三维(3D)复合材料在各工业领域的应用具有重要意义。
考点1:“绒经纤维”必须译为“ warp pile fabrics”不可译为“pile warp fabrics”因为“绒经纤维”的固定表达就是“warp pile fabrics” 考点2:“增强筋孔”推荐译为“reinforcement rib holes” 考点3:“模压固化”推荐译为“compression curing”或者“compression molding and curing”,二者选其一即可,保持译文一致性 考点4:“主控裂纹”必须译为“dominant crack”不可译为“primary crack”因为这种裂纹表示的是会引起纤维结构被完全破坏的裂纹,有固定术语进行表达 考点5:“三维机织整体成型技术”推荐译为“Integrated 3D Weaving Technology”
7857c
学术论文
自然科学
154
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 荔枝花期具有花穗大、花量多等特点,合理调控花朵雌雄比例是荔枝果园保质增产的关键作业环节。为实现荔枝花雌雄快速检测,减少人工统计误差并满足药肥精准调控需求,该文提出一种适合于部署到低功耗嵌入式平台的荔枝花雌雄智能检测方法。该方法采用Multi-Teacher Pre-Activation Features Distillation (MPFD)架构,选用结构相对复杂的YOLOv4、YOLOv5-l作为教师模型,并选用结构相对简单的YOLOv4-Tiny作为学生模型,通过动态学习不同教师模型的中间特征知识,使得学生模型在满足低功耗和实时性等嵌入式平台应用需求的基础上进一步提升其检测性能。该文的主要工作包括:优化激活函数前的蒸馏位置以降低特征蒸馏损失,使用LogCosh-Squared函数作为蒸馏距离损失函数以提高蒸馏性能,采用Margin-Activation的方法使教师模型的更多有效特征传递给学生模型,同时提出采用Conv-GN结构对学生模型进行特征转换以防止有效信息丢失。将蒸馏后的学生模型量化并移植部署到FPGA嵌入式平台,设计实现了荔枝花雌雄快速智能检测系统。实验结果表明,与未经蒸馏的学生模型相比,经过MPFD特征蒸馏后得到的学生模型mAP值提升了4.42%,达到94.21%;移植部署到FPGA嵌入式平台的检测模型大小为5.91 MB,功耗仅为10 W,较服务器平台的检测模型分别降低了73.85%和94.54%,能够更好满足荔枝花雌雄快速检测与精准统计的应用需求。 荔枝是亚热带地区广泛栽培的特产果树,因其营养丰富、口感好而具有高商业价值[1-2]。荔枝花期较短,具有花穗大、花量多等特征,并且在同一花序中,雌雄花朵异熟异开,雌花比例相对较低。因此,荔枝果园管理要监测花期信息并调控花穗花量,通过药、水、肥合理施用来提高雌花比例,为授粉受精创造有利条件,以提升荔枝坐果率[3-4]。其中,如何快速、精准采集花期信息是荔枝果园科学管理需要解决的必要技术问题。 近年来,针对荔枝果园花期信息采集与花朵检测等问题,国内外研究人员相继提出了多种基于机器视觉技术的解决方法。针对复杂自然环境下稠密聚集的荔枝花检测需求,熊俊涛等[5]提出了一种检测荔枝花、叶像素的深度语义分割网络,实现了自然环境下的荔枝花、叶等目标分割。Lin等[6]利用无人机(UAV)图像与计算机视觉技术,结合YOLOv4模型和方程拟合的方式,构建了单棵荔枝树上花簇和花丛数量的估测模型,实现了自然环境下荔枝花簇和花丛的数量自动统计。为了更好的统计荔枝花期开花强度以实现产量估计,Lin等[7]将荔枝花数及其密度图进行组合,通过使用雄性荔枝花的图像作为多列卷积神经网络的输入,生成荔枝花的密度图和数量,实现了一种优于目标检测的荔枝雄花计数方法。Ye等[8]提出了一种聚集损失和分割损失相结合的聚合损失函数,提高了检测密集花朵的鲁棒性,实现了密集环境下的荔枝花朵识别。针对其他果园花期信息采集需求,Ambrozio等[9]和Sun等[10]为了检测自然环境下落叶果树的花朵,分别应用语义分割方法,提出了端到端的残留卷积神经网络和微调后)的DeepLab-ResNet网络,实现了面向苹果、梨和桃等多场景花朵的自动检测。Wu等[11]结合通道剪枝方法提出改进YOLOv4模型,并将其应用于自然环境下不同果树品种和光照方向的苹果花检测。Dorj等[12]和Lyu等[13]分别应用颜色检测算法和级联融合的目标检测方法,设计了能够在自然环境下自动检测柑橘花的嵌入式系统。 当前的相关研究验证了自然环境下应用机器视觉技术采集荔枝花朵、果实及病虫害等信息的可行性[14-16]。然而, 荔枝花期信息统计工作存在一定困难。一方面,荔枝花量多,单支花穗的花朵数量可达数千朵,花朵目标小且相互遮挡[17-18];另一方面,受品种、树龄、气候等条件影响,不同荔枝果园的花穗花量存在显著差异,并且荔枝花期具有多批次开花的特点 [19-20]。因此,仅对荔枝花期内单一时刻的花穗花量进行检测缺少统计学意义。一般地,园艺专家会选取若干典型花穗,并使用高密度尼龙网进行套袋,待花朵全部自然脱落后,再对荔枝干花进行人工统计。另外, 为提升统计效率,收集到的荔枝花通常先人工区分为雌、雄集合,再采用称重的方式估算花量和雌雄比例,但这不可避免将导致统计数据误差。因此,研究如何实现荔枝花快速检测与精准统计更加符合荔枝果园花期信息采集的实际应用需求。
荔枝花期具有花穗大、花量多等特点,合理调控花朵雌雄比例是荔枝果园保质增产的关键作业环节。为实现荔枝花雌雄快速检测,减少人工统计误差并满足药肥精准调控需求,该文提出一种适合于部署到低功耗嵌入式平台的荔枝花雌雄智能检测方法。该方法采用Multi-Teacher Pre-Activation Features Distillation (MPFD)架构,选用结构相对复杂的YOLOv4、YOLOv5-l作为教师模型,并选用结构相对简单的YOLOv4-Tiny作为学生模型,通过动态学习不同教师模型的中间特征知识,使得学生模型在满足低功耗和实时性等嵌入式平台应用需求的基础上进一步提升其检测性能。该文的主要工作包括:优化激活函数前的蒸馏位置以降低特征蒸馏损失,使用LogCosh-Squared函数作为蒸馏距离损失函数以提高蒸馏性能,采用Margin-Activation的方法使教师模型的更多有效特征传递给学生模型,同时提出采用Conv-GN结构对学生模型进行特征转换以防止有效信息丢失。将蒸馏后的学生模型量化并移植部署到FPGA嵌入式平台,设计实现了荔枝花雌雄快速智能检测系统。实验结果表明,与未经蒸馏的学生模型相比,经过MPFD特征蒸馏后得到的学生模型mAP值提升了4.42%,达到94.21%;移植部署到FPGA嵌入式平台的检测模型大小为5.91 MB,功耗仅为10 W,较服务器平台的检测模型分别降低了73.85%和94.54%,能够更好满足荔枝花雌雄快速检测与精准统计的应用需求。 荔枝是亚热带地区广泛栽培的特产果树,因其营养丰富、口感好而具有高商业价值[1-2]。荔枝花期较短,具有花穗大、花量多等特征,并且在同一花序中,雌雄花朵异熟异开,雌花比例相对较低。因此,荔枝果园管理要监测花期信息并调控花穗花量,通过药、水、肥合理施用来提高雌花比例,为授粉受精创造有利条件,以提升荔枝坐果率[3-4]。其中,如何快速、精准采集花期信息是荔枝果园科学管理需要解决的必要技术问题。 近年来,针对荔枝果园花期信息采集与花朵检测等问题,国内外研究人员相继提出了多种基于机器视觉技术的解决方法。针对复杂自然环境下稠密聚集的荔枝花检测需求,熊俊涛等[5]提出了一种检测荔枝花、叶像素的深度语义分割网络,实现了自然环境下的荔枝花、叶等目标分割。Lin等[6]利用无人机(UAV)图像与计算机视觉技术,结合YOLOv4模型和方程拟合的方式,构建了单棵荔枝树上花簇和花丛数量的估测模型,实现了自然环境下荔枝花簇和花丛的数量自动统计。为了更好的统计荔枝花期开花强度以实现产量估计,Lin等[7]将荔枝花数及其密度图进行组合,通过使用雄性荔枝花的图像作为多列卷积神经网络的输入,生成荔枝花的密度图和数量,实现了一种优于目标检测的荔枝雄花计数方法。Ye等[8]提出了一种聚集损失和分割损失相结合的聚合损失函数,提高了检测密集花朵的鲁棒性,实现了密集环境下的荔枝花朵识别。针对其他果园花期信息采集需求,Ambrozio等[9]和Sun等[10]为了检测自然环境下落叶果树的花朵,分别应用语义分割方法,提出了端到端的残留卷积神经网络和微调后)的DeepLab-ResNet网络,实现了面向苹果、梨和桃等多场景花朵的自动检测。Wu等[11]结合通道剪枝方法提出改进YOLOv4模型,并将其应用于自然环境下不同果树品种和光照方向的苹果花检测。Dorj等[12]和Lyu等[13]分别应用颜色检测算法和级联融合的目标检测方法,设计了能够在自然环境下自动检测柑橘花的嵌入式系统。 当前的相关研究验证了自然环境下应用机器视觉技术采集荔枝花朵、果实及病虫害等信息的可行性[14-16]。然而, 荔枝花期信息统计工作存在一定困难。一方面,荔枝花量多,单支花穗的花朵数量可达数千朵,花朵目标小且相互遮挡[17-18];另一方面,受品种、树龄、气候等条件影响,不同荔枝果园的花穗花量存在显著差异,并且荔枝花期具有多批次开花的特点 [19-20]。因此,仅对荔枝花期内单一时刻的花穗花量进行检测缺少统计学意义。一般地,园艺专家会选取若干典型花穗,并使用高密度尼龙网进行套袋,待花朵全部自然脱落后,再对荔枝干花进行人工统计。另外, 为提升统计效率,收集到的荔枝花通常先人工区分为雌、雄集合,再采用称重的方式估算花量和雌雄比例,但这不可避免将导致统计数据误差。因此,研究如何实现荔枝花快速检测与精准统计更加符合荔枝果园花期信息采集的实际应用需求。
考点1:“花穗” 推荐译为 panicle(植物学术语,荔枝花为圆锥花序),避免使用不准确的 “flower spike”。 考点2:“移植部署” 必须译为 port and deploy(在技术语境中同时指代码/模型移植并部署运行)。 考点3:“人工统计误差” 必须译为 manual counting error(s),在涉及手工计数场景时优先用 counting,避免笼统用 statistical。 考点4:“模型量化” 推荐译为 model quantization(名词)或 quantize the model(动词),根据句式选择。 考点5:“病虫害” 必须译为 diseases and pests,避免遗漏“diseases”或误译为“insects”。 考点6:“合理调控” 必须译为 proper regulation 或 effective regulation,避免直译为 “reasonable regulation”。 考点7:“合理施用” 必须译为 proper application 或 appropriate application,避免直译为 “reasonable application”。
78e92
学术论文
自然科学
117
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: 1.5.1 Boeing 787 Dreamliner The Boeing 787 Dreamliner (Fig. 1.5A) is a family of long-range, midsize wide-body, twin-engine jet airliners that can seat 242–335 passengers in a typical three-class seating configuration. This aircraft, the world’s first major commercial airliner to use composite materials as the primary material in its airframe, is Boeing’s most fuel-efficient airliner [11]. The Boeing 787 maiden flight took place on December 15, 2009 and completed flight testing in mid-2011. Final Federal Aviation Administration (FAA) and European Aviation Safety Agency (EASA) type certification was received in August 2011 and the first 787-8 model was delivered to All Nippon Airways in September 2011. The Boeing 787 aircraft is 80% composite by volume. By weight, the material contents is 50% composite, 20% aluminum, 15% titanium, 10% steel, and 5% other [11]. Aluminum is used for the wing and tail leading edges; titanium is used mainly on engines and fasteners, with steel used in various areas. Each Boeing 787 aircraft contains approximately 32,000 kg of CFRP composites, made with 23 t of carbon fiber [11]. Composites are used on fuselage, wings, tail, doors, and interior. Boeing 787 fuselage sections are laid up on huge rotating mandrels (Fig. 1.6A). AFP and ATL robotic heads robotically layers of carbon-fiber epoxy resin prepreg to contoured surfaces. Reinforcing fibers are oriented in specific directions to deliver maximum strength along maximum load paths. The fuselage sections are cured in huge autoclaves. The resulting monocoque shell has internal longitudinal stiffeners already built in (Fig. 1.6B and C). This highly integrated structure requires orders of magnitude less fasteners than the conventional built-up airframes. Similar composite manufacturing techniques are applied to the wings. 5.5.3 Technical contribution to performance Air-vehicle performance is technology-based, and as technology evolves so performance is improved. This links to a gradual change in airliner design criteria, and while aircraft configuration has been relatively static, the shape of airliners has changed subtly in recent decades. It could change remarkably over coming years as new performance demands are made by customers. A new structural material having the most influence is carbon-fibre reinforced plastic (CFRP). The Airbus A380 uses some all-CFRP components and has a large proportion of the fuselage manufactured using a unique aluminium/reinforced-plastic sandwich (trade name ‘Glare’), while the newer Boeing 787 Dreamliner and its even newer Airbus equivalent aircraft, the A350, will use carbon-fibre extensively. The Boeing design is in a more advanced state of development and is anticipated to have an almost all-CFRP fuselage shell and CFRP wings. Airbus A350 plans are based on a similar technological basis. CFRP offers durability and is relatively light. The designers have to exploit the potential weight saving with care. For the 787 Dreamliner Boeing has allied airline requests for fuel efficiency improvements with a desire stemming from other areas of concern in the airports and airspace environments to offer more flexibility in operations. This is an early indication of the needs of the air transport system as a whole taking a part in the debates that range outside their direct area of interest. Boeing believe that the 787 will offer better fuel efficiency over a wide combination of payload and range values, and they also believe that being a smaller aircraft than most current long-haul types it will offer the opportunity for more point-to-point operations. The analysis of route networks in Chapter 1 has shown that airline hubbing is often the reason that some airports become congested, while others see their direct-route possibilities diminished. The 787 is being marketed as an aircraft capable of making hubs a thing of the past and opening more direct routes. The converse solution is to offer more capacity per movement at hubs, which is where the 600-800-seat A380 will be exploited. The reality will be, surely, that both types will make their mark. Where the 787 is particularly significant is that the design is using the new technology adroitly. The aircraft wing is very flexible and in its detail it is indicative of a push to achieve a higher aerodynamic efficiency (a higher lift- to-drag ratio). This will reduce fuel-burn, and will be improved by using engines that are achieving better fuel-burn performance than earlier generation engines. The combination of aerodynamic and propulsion system improvements should set a new benchmark, and on preliminary performance data released by Boeing (at September 2005), the fuel efficiency at the maximum payload-range point was shown to be 33% improved on the 777-200LR. There is probably some optimism in this figure. This has been achieved in one generation (there was an interval of about 15 years between the launch of 777 and 787 as projects), and exemplifies not just commitment by the manufacturers and their major suppliers to achieving significant environmental improvement but also in exploiting technology. In so doing, they do not minimise the cash-flow on their programmes and the risks they take are considerable. Most major companies involved in airliner design and suppliers such as aeroengine companies all delicately tread the lines that delineate technical and financial risk. Aircraft efficiency is measurable in many ways. Some tabulated data regarding leading features of some significant aircraft types that have been introduced over the last 40 years are presented in Table 5.5 to exemplify briefly how some intuitive, and sometimes counter-intuitive, developments have taken place. Behind each of the strides that has been made in the evolutionary tale there is always also a tale of risk management on a grand scale. Fuel fraction is a measure of the maximum fuel load as a proportion of the maximum take-off weight and is often quoted as a measure of technical efficiency. In the course of their life, short-medium-haul aircraft tend to complete more flights than long-haul aircraft and are more robust in terms of their construction. The consequence is that the fuel fraction is always lower than that of long-haul aircraft. (The differentiation is not so clear-cut, and the Boeing 757–200 and 767–400 are aircraft capable of trans- Atlantic operations and their fuel fraction is noticeably higher than other aircraft in their category.) Listing the selected types in date order emphasises the fact that fuel fraction is not tending to rise, although it was expected that it would as lighter-weight structural materials became available. The observation is that as more fuel-efficient engines have been developed, the necessary range performance has been attained with smaller and less heavy aircraft. The tendency for the key aerodynamic efficiency indicator, the wing aspect ratio, to rise with time is evident, albeit the rate of increase is slow. There are always exceptions to the rule. The A380 has a relatively low aspect ratio, which is attributable to the need to provide a given wing area but to fit into an apron stand ‘footprint’ that is 80 metres square. Hence aerodynamic efficiency has been compromised in order to make the aircraft compatible with modern airports.
1.5.1 Boeing 787 Dreamliner The Boeing 787 Dreamliner (Fig. 1.5A) is a family of long-range, midsize wide-body, twin-engine jet airliners that can seat 242–335 passengers in a typical three-class seating configuration. This aircraft, the world’s first major commercial airliner to use composite materials as the primary material in its airframe, is Boeing’s most fuel-efficient airliner [11]. The Boeing 787 maiden flight took place on December 15, 2009 and completed flight testing in mid-2011. Final Federal Aviation Administration (FAA) and European Aviation Safety Agency (EASA) type certification was received in August 2011 and the first 787-8 model was delivered to All Nippon Airways in September 2011. The Boeing 787 aircraft is 80% composite by volume. By weight, the material contents is 50% composite, 20% aluminum, 15% titanium, 10% steel, and 5% other [11]. Aluminum is used for the wing and tail leading edges; titanium is used mainly on engines and fasteners, with steel used in various areas. Each Boeing 787 aircraft contains approximately 32,000 kg of CFRP composites, made with 23 t of carbon fiber [11]. Composites are used on fuselage, wings, tail, doors, and interior. Boeing 787 fuselage sections are laid up on huge rotating mandrels (Fig. 1.6A). AFP and ATL robotic heads robotically layers of carbon-fiber epoxy resin prepreg to contoured surfaces. Reinforcing fibers are oriented in specific directions to deliver maximum strength along maximum load paths. The fuselage sections are cured in huge autoclaves. The resulting monocoque shell has internal longitudinal stiffeners already built in (Fig. 1.6B and C). This highly integrated structure requires orders of magnitude less fasteners than the conventional built-up airframes. Similar composite manufacturing techniques are applied to the wings. 5.5.3 Technical contribution to performance Air-vehicle performance is technology-based, and as technology evolves so performance is improved. This links to a gradual change in airliner design criteria, and while aircraft configuration has been relatively static, the shape of airliners has changed subtly in recent decades. It could change remarkably over coming years as new performance demands are made by customers. A new structural material having the most influence is carbon-fibre reinforced plastic (CFRP). The Airbus A380 uses some all-CFRP components and has a large proportion of the fuselage manufactured using a unique aluminium/reinforced-plastic sandwich (trade name ‘Glare’), while the newer Boeing 787 Dreamliner and its even newer Airbus equivalent aircraft, the A350, will use carbon-fibre extensively. The Boeing design is in a more advanced state of development and is anticipated to have an almost all-CFRP fuselage shell and CFRP wings. Airbus A350 plans are based on a similar technological basis. CFRP offers durability and is relatively light. The designers have to exploit the potential weight saving with care. For the 787 Dreamliner Boeing has allied airline requests for fuel efficiency improvements with a desire stemming from other areas of concern in the airports and airspace environments to offer more flexibility in operations. This is an early indication of the needs of the air transport system as a whole taking a part in the debates that range outside their direct area of interest. Boeing believe that the 787 will offer better fuel efficiency over a wide combination of payload and range values, and they also believe that being a smaller aircraft than most current long-haul types it will offer the opportunity for more point-to-point operations. The analysis of route networks in Chapter 1 has shown that airline hubbing is often the reason that some airports become congested, while others see their direct-route possibilities diminished. The 787 is being marketed as an aircraft capable of making hubs a thing of the past and opening more direct routes. The converse solution is to offer more capacity per movement at hubs, which is where the 600-800-seat A380 will be exploited. The reality will be, surely, that both types will make their mark. Where the 787 is particularly significant is that the design is using the new technology adroitly. The aircraft wing is very flexible and in its detail it is indicative of a push to achieve a higher aerodynamic efficiency (a higher lift- to-drag ratio). This will reduce fuel-burn, and will be improved by using engines that are achieving better fuel-burn performance than earlier generation engines. The combination of aerodynamic and propulsion system improvements should set a new benchmark, and on preliminary performance data released by Boeing (at September 2005), the fuel efficiency at the maximum payload-range point was shown to be 33% improved on the 777-200LR. There is probably some optimism in this figure. This has been achieved in one generation (there was an interval of about 15 years between the launch of 777 and 787 as projects), and exemplifies not just commitment by the manufacturers and their major suppliers to achieving significant environmental improvement but also in exploiting technology. In so doing, they do not minimise the cash-flow on their programmes and the risks they take are considerable. Most major companies involved in airliner design and suppliers such as aeroengine companies all delicately tread the lines that delineate technical and financial risk. Aircraft efficiency is measurable in many ways. Some tabulated data regarding leading features of some significant aircraft types that have been introduced over the last 40 years are presented in Table 5.5 to exemplify briefly how some intuitive, and sometimes counter-intuitive, developments have taken place. Behind each of the strides that has been made in the evolutionary tale there is always also a tale of risk management on a grand scale. Fuel fraction is a measure of the maximum fuel load as a proportion of the maximum take-off weight and is often quoted as a measure of technical efficiency. In the course of their life, short-medium-haul aircraft tend to complete more flights than long-haul aircraft and are more robust in terms of their construction. The consequence is that the fuel fraction is always lower than that of long-haul aircraft. (The differentiation is not so clear-cut, and the Boeing 757–200 and 767–400 are aircraft capable of trans- Atlantic operations and their fuel fraction is noticeably higher than other aircraft in their category.) Listing the selected types in date order emphasises the fact that fuel fraction is not tending to rise, although it was expected that it would as lighter-weight structural materials became available. The observation is that as more fuel-efficient engines have been developed, the necessary range performance has been attained with smaller and less heavy aircraft. The tendency for the key aerodynamic efficiency indicator, the wing aspect ratio, to rise with time is evident, albeit the rate of increase is slow. There are always exceptions to the rule. The A380 has a relatively low aspect ratio, which is attributable to the need to provide a given wing area but to fit into an apron stand ‘footprint’ that is 80 metres square. Hence aerodynamic efficiency has been compromised in order to make the aircraft compatible with modern airports.
考点1:“All Nippon Airways”需译为“全日空航空”,译为“全日本航空”算错误 考点2:“titanium”推荐译为“钛” 考点3:“robotically”推荐译为“自动” 考点4:“longitudinal stiffeners”推荐译为“纵向加劲肋” 考点5:“The designers have to exploit the potential weight saving with care”推荐译为“设计师在挖掘其减重潜力时,必须十分谨慎。” 考点6:“payload”推荐译为“有效载荷” 考点7:“making hubs a thing of the past”推荐译为“打破枢纽束缚” 考点8:“flexible”推荐译为“柔韧”
797fa
学术论文
应用学科
9
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务:  The city of Rome had sustained little diminution in her architectural splendour, when the setting sun shed his parting beams, as if with prophetic significance, upon the gilded roof of the venerable capitol, on the evening of the 24th of August, A.D. 410. The temple of Jupiter, though shorn of some of the dazzling ornaments with which the emperor Domitian had adorned its portals and pediments, still remained an imposing monument of the ancient paganism of the Imperial City. Other costly temples and public buildings, clustered around that seat of Roman pride and greatness, and met and charmed the eye of the citizen, as he ascended the slope of the Capitoline Hill. With a lordly air, these noble structures threw their long shadows over the spacious forum, where, of old, the sons of the republic had been accustomed to gather in crowds around the rostrum, to listen to the speeches of their orators; and where still the degenerate Roman was reminded of the deeds of his fathers, by the monuments of patriotism and victory which were strewn around him. On that evening, might be seen many a citizen and foreigner passing to and fro, along its stately colonnades, or reclining at his ease upon the marble seats; and in whatever direction he went, on leaving that far-famed spot, he passed through squares and streets which were adorned with temples, palaces, and baths, such as could have been erected in no city but one that had enriched itself with the spoils of the whole world. In short, Rome had undergone but little alteration since the eastern emperor Constantius, fifty years before, on visiting the city of his fathers, had been overwhelmed with astonishment at its surpassing magnificence. An historian of that period,[1] describing the visit in that inflated style which is so characteristic of the age, observes; "As Constantius viewed the seven-hilled city, with its valleys and suburban districts, every object around him seemed to shine with transcendent splendour:—the temple of Tarpeian Jove exceeding everything he had beheld, as much as a Divine production could exceed the works of man; the spacious baths spreading around like provinces; the Amphitheatre with its solid walls of Tiburtine marble, and so lofty, that the eye is fatigued in looking upward to its summit; the Pantheon with its vast circular space, arched over by a magnificent dome; and its lofty pediments rising one above another, and crowned with statues of Roman heroes; the Forum and Temple of Peace; the Theatre of Pompey; the Musical Hall; the Stadia, and other imposing objects in the Eternal City. But when he came to the Forum of Trajan—the most astonishing structure under the face of heaven, and, as I conceive, wonderful in the estimation of the deities themselves—he was struck with astonishment, while considering its gigantic buildings, which are not to be described in language, or again to be equalled by mortal skill. Discarding the idea of erecting another forum like that, he thought that he might rear an equestrian statue, which should resemble the colossal horse of Trajan; but this design he also abandoned, upon hearing it remarked by the prince, Hormisdas, 'If you would succeed in having a similar horse, you must first provide a similar stable,'" Such was the grandeur of ancient Rome; and it was probably with feelings of admiration like those of the emperor and his historian, that many a citizen returned from the baths and the forums to his own dwelling on the eventful evening in question. Gradually the sounds of business, and the murmur of voices in the streets died away: and as the stars shone forth in the face of heaven, the mighty city slept in silence. But it was a silence soon to be disturbed. At the midnight hour, a blast of trumpets like the roar of thunder reverberated from hill to hill, and woke up myriads of the inhabitants from their deep slumbers—it was the signal that Alaric the Goth, with his mighty army, had entered Rome. Two years before the barbarian general had besieged the city. Swayed by what he conceived a supernatural impulse, he led his victorious troops down the passes of the Apennines, upon the rich plains of Italy. A pious monk, it is said, met the warrior on his way, and exhorted him to refrain from his expedition; but he replied, "I am urged on in spite of myself, by an irresistible impulse which is continually saving to me, 'March to Rome, and desolate the city.'"[2] Thus, prompted by his ambition, he fulfilled his destiny, and wreaked a fearful amount of vengeance on the heads of the Romans, for the wrongs which they had inflicted upon others. Twice did he blockade the gates of Rome, and subdue the proud masters of the world. During the first siege, the terrors of famine and pestilence reduced the senate to submission, and the conqueror agreed to raise the siege, only upon the condition of his being paid a very large ransom. Negotiations for peace with the emperor Honorius, who was then at Ravenna, having failed, Alaric returned to Rome, and again pitched his camp before the walls. The remembrance of their calamities, during the former siege, constrained the people once more to yield; when the Gothic warrior insisted upon their renouncing allegiance to Honorius, and imposed upon them a new emperor in the person of Attalus, the prefect of the city. But it was not long before the latter forfeited the confidence of his master, and Alaric immediately proceeded publicly to strip him of the imperial purple. The Goth, after this circumstance, renewed his negotiations with the court of Ravenna; but being insulted by the heralds, and attacked by the troops of Honorius, he turned his army a third time towards the gates of Rome.[3] Historians inform us, that it was by an act of treachery, that Alaric was now admitted into the city; but no satisfactory information can be obtained respecting the particulars of the important transaction. The Gothic trumpet, however, at the Salarian gate, the march of the enemy along the great highway, and the flames issuing from the palace of Sallust—which was fired by the troops, as soon as they entered within the walls—proclaimed that Rome, the Queen of Cities, after the lapse of nearly eight hundred years from her invasion by the Gauls, was once more in the hands of a barbarian foe. Although the Romans had been aware of the vicinity of Alaric, yet, lulled into a state of false security, they did not anticipate any assault, and the senators were quietly slumbering in their beds when the enemy entered the city. Fearful were the scenes enacted; and well might Jerome apply to it the lines of Virgil, in reference to the sack of Troy: "What tongue can tell the slaughter of that night? What eyes can weep the sorrow and affright? An ancient and imperial city falls— The streets are fill'd with frequent funerals; Houses and holy temples float in blood, And hostile nations made a common flood; All parts resound with tumults, plaints, and fears, Ana grisly death in sundry shapes appears." The cruel and licentious soldiery made a dreadful slaughter of the Roman people, and violated many a matron and virgin. The horrors of the invasion were further heightened by the excesses which were practised by forty thousand slaves, who now broke loose from the authority of their masters, and retaliated, on them and their families, the wrongs which themselves and their predecessors had endured through ages of oppression. But it is acknowledged by all writers, that Alaric—who was himself an Arian—showed some considerable regard for the Christians of the city, and spared the churches where they met for worship. Indeed he appointed the edifices, which had been dedicated to the apostles Peter and Paul, as places of refuge for the terrified Christian inhabitants, and gave strict orders that those who fled there for sanctuary, should be protected from injury. Instances illustrative of the forbearance of the soldiers, and of their respect not only for the persons of the Christians but for the consecrated vessels which they employed in their worship, are afforded us by the historians of those times. Orosius gives us a graphic description of a long train of Christians, carrying on their heads the communion-plate of gold and silver, and singing their sacred hymns, who were escorted in safety, by the Gothic soldiers, through the streets of the ravaged city, to the church of St. Peter. He speaks also of many of the barbarians, and the pagan Romans, uniting in these songs, and joining in the solemn procession; and represents the latter as saving themselves from vengeance, by taking shelter beneath the wing of the Christian faith.
 The city of Rome had sustained little diminution in her architectural splendour, when the setting sun shed his parting beams, as if with prophetic significance, upon the gilded roof of the venerable capitol, on the evening of the 24th of August, A.D. 410. The temple of Jupiter, though shorn of some of the dazzling ornaments with which the emperor Domitian had adorned its portals and pediments, still remained an imposing monument of the ancient paganism of the Imperial City. Other costly temples and public buildings, clustered around that seat of Roman pride and greatness, and met and charmed the eye of the citizen, as he ascended the slope of the Capitoline Hill. With a lordly air, these noble structures threw their long shadows over the spacious forum, where, of old, the sons of the republic had been accustomed to gather in crowds around the rostrum, to listen to the speeches of their orators; and where still the degenerate Roman was reminded of the deeds of his fathers, by the monuments of patriotism and victory which were strewn around him. On that evening, might be seen many a citizen and foreigner passing to and fro, along its stately colonnades, or reclining at his ease upon the marble seats; and in whatever direction he went, on leaving that far-famed spot, he passed through squares and streets which were adorned with temples, palaces, and baths, such as could have been erected in no city but one that had enriched itself with the spoils of the whole world. In short, Rome had undergone but little alteration since the eastern emperor Constantius, fifty years before, on visiting the city of his fathers, had been overwhelmed with astonishment at its surpassing magnificence. An historian of that period,[1] describing the visit in that inflated style which is so characteristic of the age, observes; "As Constantius viewed the seven-hilled city, with its valleys and suburban districts, every object around him seemed to shine with transcendent splendour:—the temple of Tarpeian Jove exceeding everything he had beheld, as much as a Divine production could exceed the works of man; the spacious baths spreading around like provinces; the Amphitheatre with its solid walls of Tiburtine marble, and so lofty, that the eye is fatigued in looking upward to its summit; the Pantheon with its vast circular space, arched over by a magnificent dome; and its lofty pediments rising one above another, and crowned with statues of Roman heroes; the Forum and Temple of Peace; the Theatre of Pompey; the Musical Hall; the Stadia, and other imposing objects in the Eternal City. But when he came to the Forum of Trajan—the most astonishing structure under the face of heaven, and, as I conceive, wonderful in the estimation of the deities themselves—he was struck with astonishment, while considering its gigantic buildings, which are not to be described in language, or again to be equalled by mortal skill. Discarding the idea of erecting another forum like that, he thought that he might rear an equestrian statue, which should resemble the colossal horse of Trajan; but this design he also abandoned, upon hearing it remarked by the prince, Hormisdas, 'If you would succeed in having a similar horse, you must first provide a similar stable,'" Such was the grandeur of ancient Rome; and it was probably with feelings of admiration like those of the emperor and his historian, that many a citizen returned from the baths and the forums to his own dwelling on the eventful evening in question. Gradually the sounds of business, and the murmur of voices in the streets died away: and as the stars shone forth in the face of heaven, the mighty city slept in silence. But it was a silence soon to be disturbed. At the midnight hour, a blast of trumpets like the roar of thunder reverberated from hill to hill, and woke up myriads of the inhabitants from their deep slumbers—it was the signal that Alaric the Goth, with his mighty army, had entered Rome. Two years before the barbarian general had besieged the city. Swayed by what he conceived a supernatural impulse, he led his victorious troops down the passes of the Apennines, upon the rich plains of Italy. A pious monk, it is said, met the warrior on his way, and exhorted him to refrain from his expedition; but he replied, "I am urged on in spite of myself, by an irresistible impulse which is continually saving to me, 'March to Rome, and desolate the city.'"[2] Thus, prompted by his ambition, he fulfilled his destiny, and wreaked a fearful amount of vengeance on the heads of the Romans, for the wrongs which they had inflicted upon others. Twice did he blockade the gates of Rome, and subdue the proud masters of the world. During the first siege, the terrors of famine and pestilence reduced the senate to submission, and the conqueror agreed to raise the siege, only upon the condition of his being paid a very large ransom. Negotiations for peace with the emperor Honorius, who was then at Ravenna, having failed, Alaric returned to Rome, and again pitched his camp before the walls. The remembrance of their calamities, during the former siege, constrained the people once more to yield; when the Gothic warrior insisted upon their renouncing allegiance to Honorius, and imposed upon them a new emperor in the person of Attalus, the prefect of the city. But it was not long before the latter forfeited the confidence of his master, and Alaric immediately proceeded publicly to strip him of the imperial purple. The Goth, after this circumstance, renewed his negotiations with the court of Ravenna; but being insulted by the heralds, and attacked by the troops of Honorius, he turned his army a third time towards the gates of Rome.[3] Historians inform us, that it was by an act of treachery, that Alaric was now admitted into the city; but no satisfactory information can be obtained respecting the particulars of the important transaction. The Gothic trumpet, however, at the Salarian gate, the march of the enemy along the great highway, and the flames issuing from the palace of Sallust—which was fired by the troops, as soon as they entered within the walls—proclaimed that Rome, the Queen of Cities, after the lapse of nearly eight hundred years from her invasion by the Gauls, was once more in the hands of a barbarian foe. Although the Romans had been aware of the vicinity of Alaric, yet, lulled into a state of false security, they did not anticipate any assault, and the senators were quietly slumbering in their beds when the enemy entered the city. Fearful were the scenes enacted; and well might Jerome apply to it the lines of Virgil, in reference to the sack of Troy: "What tongue can tell the slaughter of that night? What eyes can weep the sorrow and affright? An ancient and imperial city falls— The streets are fill'd with frequent funerals; Houses and holy temples float in blood, And hostile nations made a common flood; All parts resound with tumults, plaints, and fears, Ana grisly death in sundry shapes appears." The cruel and licentious soldiery made a dreadful slaughter of the Roman people, and violated many a matron and virgin. The horrors of the invasion were further heightened by the excesses which were practised by forty thousand slaves, who now broke loose from the authority of their masters, and retaliated, on them and their families, the wrongs which themselves and their predecessors had endured through ages of oppression. But it is acknowledged by all writers, that Alaric—who was himself an Arian—showed some considerable regard for the Christians of the city, and spared the churches where they met for worship. Indeed he appointed the edifices, which had been dedicated to the apostles Peter and Paul, as places of refuge for the terrified Christian inhabitants, and gave strict orders that those who fled there for sanctuary, should be protected from injury. Instances illustrative of the forbearance of the soldiers, and of their respect not only for the persons of the Christians but for the consecrated vessels which they employed in their worship, are afforded us by the historians of those times. Orosius gives us a graphic description of a long train of Christians, carrying on their heads the communion-plate of gold and silver, and singing their sacred hymns, who were escorted in safety, by the Gothic soldiers, through the streets of the ravaged city, to the church of St. Peter. He speaks also of many of the barbarians, and the pagan Romans, uniting in these songs, and joining in the solemn procession; and represents the latter as saving themselves from vengeance, by taking shelter beneath the wing of the Christian faith.
考点1:【capitol】应译为 “卡比托利欧山/卡比托利欧” 考点2:【Constantius】 应译为 “君士坦提乌斯” 考点3:【portals and pediments】 应译为 “门廊与山花” 考点4:【sons of the republic】 应译为 “共和国的子民” 考点5:【eastern emperor】 应译为 “东罗马皇帝”
7a4a0
文学艺术
小说
78
翻译,英文翻译成中文。不要输出译文以外的内容。以下是你本次的任务: Abstract: America’s history of literature began with the swarming in of immigrants with different background and cultures. After that, American literature had been greatly influenced by the European culture for a long period. It was not until America’s independence, did Americans realized that they need national literature strongly, and American literature began to developed. The Civil War was a watershed in the history, after which American literature entered a period of full blooming. Romantics, which emphasized individualism and intuition and Tnscendentalism represented by Emerson came out into being. This was an exciting period in the history of American literature. Like the flowers of spring, there were suddenly many different kinds of writing at the same time. They have given depth and strength to American literature, and accelerated the forming of High Romantics. But due to the influence of Civil War, the American society was in a turbulent situation. The writings about local life, critical realism and unveiling the dark side of the society were increased. After The First World War, Americans were at a loss postwar, and the Modern American literature began. My piece of paper is written in chronological order as these periods developed in order to have a clear outline of its progress. American is a multi-national country. Just like a big container, which put in various kinds of elements. Different cultures, that can not only be co-existed but also form a sharp contrast, mixed together, It makes American literature style has a flavor of distinct and various aesthetic feeling. Many writers come from lower level, which makes American literature has the rich flavor of life and local color. Furthermore, many new styles of literature in the world are oriented in America since 20th century. The process of American literature can be divided into following main periods: Colony and Puritan literature; early national literature; latter national literature and Modern literature. 1. Colonial and early American literature (1). Travelers and Explorers When the European explorers first came to this new continent, the native Indians who probably got here from Asia about fifteen thousand years ago were still in origin, and they even had no written language, “The traditional literature was originally transmitted almost entirely by word of mouth, and therefore belongs to the category of oral literature,” (Wu Dingbo, 1) As time past, more and more travelers and explorers swarmed in. They wrote a lot of diaries、letters, and travel accounts to describe the new land as second Eden. No wander somebody said that the earliest American literature were the travel accounts written by European adventurers. Among the most remained were Captain John Smith’s True Relation of Virginia (1608), and Description of New England (1616). Although most of the Indian history was preserved in tales and songs, they had thoughts about life and nature. They loved the natural world around them deeply, and they believed that when a person was dead, he would give back what had borrowed while he was alive to nature. This kind of philosophy had influenced later or even modern American writers. It’s interesting that when we look at the literature of the Puritans, the Transcendentalists, the Naturalists, and even the Moderns, when we read Anne Bradstreet, Emily Dickinson, Stephen Crane, and Ernest Hemingway, we can find similar themes. (2) Pilgrim settlements Several years later, another group of settlers also arrived in the New World. This group was looking for the Jamestown settlement. However, because of bad navigation, they landed in Massachusetts. They were also coming to the New World with dreams of success, but their goal was different from the Jamestown settlement. They wanted to start a new world governed by the Bible. They were called Puritans because they wanted to live a better life by making themselves pure. They first arrived on the Mayflower and settled in Plymouth. This is the group we are usually thinking about when we talk about the "first Americans." The clearest history of their journey to the New World can be found in History of Plymouth Plantation (1608) written by William Bradford, who was also one of the Mayflower passengers. The History of Plymouth Plantation is a Puritan book in the best sense. “It’s loosely annalistic, but a direct and simple style gives charm, as a sincere faith in Puritanism gives purity, to the entire book.” (W. P. Trent, 1997) The Puritans had several kinds of literature. By far the most common form is the writing related to Biblical teachings, or sermons, that the church leaders wrote. The Puritans believed that they were in the New World because God had brought them there for a special purpose. They thought that by studying the Bible they could learn more about this way of life. So they were very strict to their life, and they didn’t allow any kind of entertainment even in literature. That’s way Wu Dingbo said in his book “Literature of the New England Settlement is mainly a literary expression of the Puritan idealism” and “The literature of the colonial settlement served either God or colonial expansion or both.” (Wu Dingbo, 4) Another important form of writing from this period is the histories. These books, like Bradford's History of Plymouth Plantation, are important because they tell us about life at the time of the Puritans. People also wrote many poems. But a lot of works were hidden and lost because people often considered poetry to be an inferior form of writing and not totally acceptable to Puritan thinking. One of the most significant poets from this period was Anne Bradstreet (1612-1672). Her poems in Tenth Muse Lately Sprung Up In America (1650) reflected the con concerns of women who came to settle in the colonies, and in all her poems, however, she shows her strong belief in God. 2. 18th century—the Age of reason (1) The Age of reason In the 18th century, people believed in man’s own nature and the power of human reason. With Franklin as its spokesman, the 18th century America experienced an age of reason.Words had never been so useful and so important in human history. People wrote a lot of political writings.
Abstract: America’s history of literature began with the swarming in of immigrants with different background and cultures. After that, American literature had been greatly influenced by the European culture for a long period. It was not until America’s independence, did Americans realized that they need national literature strongly, and American literature began to developed. The Civil War was a watershed in the history, after which American literature entered a period of full blooming. Romantics, which emphasized individualism and intuition and Tnscendentalism represented by Emerson came out into being. This was an exciting period in the history of American literature. Like the flowers of spring, there were suddenly many different kinds of writing at the same time. They have given depth and strength to American literature, and accelerated the forming of High Romantics. But due to the influence of Civil War, the American society was in a turbulent situation. The writings about local life, critical realism and unveiling the dark side of the society were increased. After The First World War, Americans were at a loss postwar, and the Modern American literature began. My piece of paper is written in chronological order as these periods developed in order to have a clear outline of its progress. American is a multi-national country. Just like a big container, which put in various kinds of elements. Different cultures, that can not only be co-existed but also form a sharp contrast, mixed together, It makes American literature style has a flavor of distinct and various aesthetic feeling. Many writers come from lower level, which makes American literature has the rich flavor of life and local color. Furthermore, many new styles of literature in the world are oriented in America since 20th century. The process of American literature can be divided into following main periods: Colony and Puritan literature; early national literature; latter national literature and Modern literature. 1. Colonial and early American literature (1). Travelers and Explorers When the European explorers first came to this new continent, the native Indians who probably got here from Asia about fifteen thousand years ago were still in origin, and they even had no written language, “The traditional literature was originally transmitted almost entirely by word of mouth, and therefore belongs to the category of oral literature,” (Wu Dingbo, 1) As time past, more and more travelers and explorers swarmed in. They wrote a lot of diaries、letters, and travel accounts to describe the new land as second Eden. No wander somebody said that the earliest American literature were the travel accounts written by European adventurers. Among the most remained were Captain John Smith’s True Relation of Virginia (1608), and Description of New England (1616). Although most of the Indian history was preserved in tales and songs, they had thoughts about life and nature. They loved the natural world around them deeply, and they believed that when a person was dead, he would give back what had borrowed while he was alive to nature. This kind of philosophy had influenced later or even modern American writers. It’s interesting that when we look at the literature of the Puritans, the Transcendentalists, the Naturalists, and even the Moderns, when we read Anne Bradstreet, Emily Dickinson, Stephen Crane, and Ernest Hemingway, we can find similar themes. (2) Pilgrim settlements Several years later, another group of settlers also arrived in the New World. This group was looking for the Jamestown settlement. However, because of bad navigation, they landed in Massachusetts. They were also coming to the New World with dreams of success, but their goal was different from the Jamestown settlement. They wanted to start a new world governed by the Bible. They were called Puritans because they wanted to live a better life by making themselves pure. They first arrived on the Mayflower and settled in Plymouth. This is the group we are usually thinking about when we talk about the "first Americans." The clearest history of their journey to the New World can be found in History of Plymouth Plantation (1608) written by William Bradford, who was also one of the Mayflower passengers. The History of Plymouth Plantation is a Puritan book in the best sense. “It’s loosely annalistic, but a direct and simple style gives charm, as a sincere faith in Puritanism gives purity, to the entire book.” (W. P. Trent, 1997) The Puritans had several kinds of literature. By far the most common form is the writing related to Biblical teachings, or sermons, that the church leaders wrote. The Puritans believed that they were in the New World because God had brought them there for a special purpose. They thought that by studying the Bible they could learn more about this way of life. So they were very strict to their life, and they didn’t allow any kind of entertainment even in literature. That’s way Wu Dingbo said in his book “Literature of the New England Settlement is mainly a literary expression of the Puritan idealism” and “The literature of the colonial settlement served either God or colonial expansion or both.” (Wu Dingbo, 4) Another important form of writing from this period is the histories. These books, like Bradford's History of Plymouth Plantation, are important because they tell us about life at the time of the Puritans. People also wrote many poems. But a lot of works were hidden and lost because people often considered poetry to be an inferior form of writing and not totally acceptable to Puritan thinking. One of the most significant poets from this period was Anne Bradstreet (1612-1672). Her poems in Tenth Muse Lately Sprung Up In America (1650) reflected the con concerns of women who came to settle in the colonies, and in all her poems, however, she shows her strong belief in God. 2. 18th century—the Age of reason (1) The Age of reason In the 18th century, people believed in man’s own nature and the power of human reason. With Franklin as its spokesman, the 18th century America experienced an age of reason.Words had never been so useful and so important in human history. People wrote a lot of political writings.
考点1:"Tenth Muse Lately Sprung Up In America"应译为《在美洲新近出现的第十位缪斯》。 考点2:"the Age of reason''应该翻译为“理性主义时代” 考点3:"swarming in"推荐译为“蜂拥而至”。 考点4:"Transcendentalism"推荐译为“超验主义”,是美国文学和哲学史上的一个专有名词。这里应该翻译为“超验主义者” 考点5: "second Eden'应译为“第二个伊甸园”。 考点6:"High Romantics"应译为“浪漫主义高潮”。 考点7:"Pilgrim settlements"应译为“清教徒定居点”
7add8
学术论文
人文科学
42
翻译,中文翻译成英文。不要输出译文以外的内容。 以下是你本次的任务: 科技日报记者 薛岩 记者从中国科协获悉,中国科协生命科学学会联合体开展的年度“中国生命科学十大进展”评选结果揭晓,7个知识创新类和3个技术创新类成果脱颖而出。 本次入选的研究项目均面向生命科学前沿,面向人民生命健康,聚焦解决热点问题,具有原创性突出、社会意义重大的特点。十大成果分别是: 一、衰老的时空编程及干预 衰老是复杂的生物学过程,具有显著时空异质性。在复杂时空背景下,揭示衰老的关键驱动因素和开发安全有效的干预措施,是衰老科学研究面临的重大挑战。中国科学院动物研究所刘光慧、分子细胞科学卓越创新中心周斌、北京基因组研究所张维绮、动物研究所曲静等团队,为衰老研究提供新范式,也为衰老及其相关疾病的临床干预奠定基础,受到国际专业期刊的亮点评价。 二、新型肠道菌源酶跨界调控代谢性疾病的作用机制与防治新策略 肠道菌群与疾病发生发展密切相关,但目前靶向菌群多为整体性干预措施,缺乏特异性治疗策略,如何精准调控肠道菌群是国际生命科学研究的前沿问题。北京大学姜长涛、乔杰及山东大学刘双江团队等合作,提出“肠道菌源酶跨界调控宿主稳态”新理论,开发点击代谢组学技术等。团队进一步围绕BAS-suc开展了工程菌构建、激动剂开发等系列临床转化研究,开辟了代谢性疾病干预新路径。 三、通用型CAR-T细胞治疗复发难治性自身免疫性疾病 长期以来,彻底治愈红斑狼疮、硬皮病、多发性硬化症等自身免疫性疾病,是全球共同面临的医学难题。现有免疫抑制药物虽然可在一定程度上缓解病情,却不能完全阻止疾病的进展,反而可能带来使人虚弱的副作用。由中国人民解放军海军军医大学长征医院徐沪济领衔、联合华东师范大学杜冰、刘明耀团队以及上海邦耀生物科技有限公司,采用我国自主研发的异体CAR-T细胞开展临床研究,在全球首次成功治愈了严重自身免疫性疾病。 四、异染色质形成机制 旁着丝粒区域由快速演化且高度重复的卫星序列组成,该区域在不同物种间均以异染色质状态存在。然而,两个长期的未解之谜是“旁着丝粒区是怎样被识别并起始异染色质形成的”以及“不同物种间不保守的卫星序列是如何被保守的分子机器所识别的”。中国科学院生物物理研究所朱冰团队发现,保守的锌指蛋白ZNF512利用其非连续排布的锌指结构域识别富集于旁着丝粒区内的保守的不连续的TTC序列元件,并直接招募SUV39H家族蛋白,从而起始异染色质形成。在回答了近百年的科学困惑的同时,本发现也为TTC元件异常扩增导致FXN基因异染色质化并诱发共济失调症提供了机制性解释。 五、紫杉醇生物合成 紫杉醇是全球销量第一的植物天然广谱抗癌药物。然而,紫杉醇生产高度依赖红豆杉资源和化学合成,环境污染大,生产成本高。中国农业科学院农业基因组研究所闫建斌团队联合北京大学雷晓光等团队,发现了紫杉醇生物合成途径中的关键酶,阐明了颠覆传统认知的紫杉醇关键药效基团的形成机制。该成果被认为是开启紫杉醇生物制造领域大门的重要里程碑,标志着“中国在紫杉醇生物合成理论和技术上处于世界领先地位”。 六、实现一年生与多年生植物的自由转换 培育和推广多年生作物对于推动农业可持续性发展和应对未来气候变化具有重要意义。然而,由于缺乏研究模式,实验周期长,“多年生特性是由什么基因来决定的”一直是一个难以破解的“黑匣子”。中国科学院分子植物科学卓越中心王佳伟研究组通过构建跨物种遗传群体和正向遗传学手段,成功定位了控制植物多年生生活习性的三个关键MADS-box基因。该研究首次克隆了决定植物寿命长短的关键基因,并实现了一年生与多年生植物的自由转换,相关成果为设计与培育多年生作物奠定了理论基础,并入选相关国际专业期刊2024年度最佳论文之一。
科技日报记者 薛岩 记者从中国科协获悉,中国科协生命科学学会联合体开展的年度“中国生命科学十大进展”评选结果揭晓,7个知识创新类和3个技术创新类成果脱颖而出。 本次入选的研究项目均面向生命科学前沿,面向人民生命健康,聚焦解决热点问题,具有原创性突出、社会意义重大的特点。十大成果分别是: 一、衰老的时空编程及干预 衰老是复杂的生物学过程,具有显著时空异质性。在复杂时空背景下,揭示衰老的关键驱动因素和开发安全有效的干预措施,是衰老科学研究面临的重大挑战。中国科学院动物研究所刘光慧、分子细胞科学卓越创新中心周斌、北京基因组研究所张维绮、动物研究所曲静等团队,为衰老研究提供新范式,也为衰老及其相关疾病的临床干预奠定基础,受到国际专业期刊的亮点评价。 二、新型肠道菌源酶跨界调控代谢性疾病的作用机制与防治新策略 肠道菌群与疾病发生发展密切相关,但目前靶向菌群多为整体性干预措施,缺乏特异性治疗策略,如何精准调控肠道菌群是国际生命科学研究的前沿问题。北京大学姜长涛、乔杰及山东大学刘双江团队等合作,提出“肠道菌源酶跨界调控宿主稳态”新理论,开发点击代谢组学技术等。团队进一步围绕BAS-suc开展了工程菌构建、激动剂开发等系列临床转化研究,开辟了代谢性疾病干预新路径。 三、通用型CAR-T细胞治疗复发难治性自身免疫性疾病 长期以来,彻底治愈红斑狼疮、硬皮病、多发性硬化症等自身免疫性疾病,是全球共同面临的医学难题。现有免疫抑制药物虽然可在一定程度上缓解病情,却不能完全阻止疾病的进展,反而可能带来使人虚弱的副作用。由中国人民解放军海军军医大学长征医院徐沪济领衔、联合华东师范大学杜冰、刘明耀团队以及上海邦耀生物科技有限公司,采用我国自主研发的异体CAR-T细胞开展临床研究,在全球首次成功治愈了严重自身免疫性疾病。 四、异染色质形成机制 旁着丝粒区域由快速演化且高度重复的卫星序列组成,该区域在不同物种间均以异染色质状态存在。然而,两个长期的未解之谜是“旁着丝粒区是怎样被识别并起始异染色质形成的”以及“不同物种间不保守的卫星序列是如何被保守的分子机器所识别的”。中国科学院生物物理研究所朱冰团队发现,保守的锌指蛋白ZNF512利用其非连续排布的锌指结构域识别富集于旁着丝粒区内的保守的不连续的TTC序列元件,并直接招募SUV39H家族蛋白,从而起始异染色质形成。在回答了近百年的科学困惑的同时,本发现也为TTC元件异常扩增导致FXN基因异染色质化并诱发共济失调症提供了机制性解释。 五、紫杉醇生物合成 紫杉醇是全球销量第一的植物天然广谱抗癌药物。然而,紫杉醇生产高度依赖红豆杉资源和化学合成,环境污染大,生产成本高。中国农业科学院农业基因组研究所闫建斌团队联合北京大学雷晓光等团队,发现了紫杉醇生物合成途径中的关键酶,阐明了颠覆传统认知的紫杉醇关键药效基团的形成机制。该成果被认为是开启紫杉醇生物制造领域大门的重要里程碑,标志着“中国在紫杉醇生物合成理论和技术上处于世界领先地位”。 六、实现一年生与多年生植物的自由转换 培育和推广多年生作物对于推动农业可持续性发展和应对未来气候变化具有重要意义。然而,由于缺乏研究模式,实验周期长,“多年生特性是由什么基因来决定的”一直是一个难以破解的“黑匣子”。中国科学院分子植物科学卓越中心王佳伟研究组通过构建跨物种遗传群体和正向遗传学手段,成功定位了控制植物多年生生活习性的三个关键MADS-box基因。该研究首次克隆了决定植物寿命长短的关键基因,并实现了一年生与多年生植物的自由转换,相关成果为设计与培育多年生作物奠定了理论基础,并入选相关国际专业期刊2024年度最佳论文之一。
考点1:中国科协生命科学学会联合体 应译为 China Association for Science and Technology Life Sciences Society Consortium 考点2:中国生命科学十大进展 应译为 Top 10 Advances in Life Sciences in China 考点3:衰老的时空编程及干预 应译为 Spatiotemporal Programming and Intervention of Aging 考点4:衰老科学研究 应译为 aging science research 考点5:新型肠道菌源酶跨界调控代谢性疾病的作用机制与防治新策略 应译为 Mechanism and New Strategies of Cross-Regulation of Metabolic Diseases by Novel Gut-Derived Enzymes 考点6:肠道菌群 应译为 gut microbiota 考点7:宿主稳态 应译为 host homeostasis 考点8:点击代谢组学 应译为 click metabolomics 考点9:通用型CAR-T细胞治疗复发难治性自身免疫性疾病 应译为 Universal CAR-T Cell Therapy for Refractory and Relapsing Autoimmune Diseases 考点10:红斑狼疮 应译为 systemic lupus erythematosus 考点11:硬皮病 应译为 scleroderma 考点12:多发性硬化症 应译为 multiple sclerosis 考点13:异染色质形成机制 应译为 Mechanism of Heterochromatin Formation 考点14:旁着丝粒区域 应译为 pericentromeric region 考点15:卫星序列 应译为 satellite sequences 考点16:锌指蛋白ZNF512 应译为 zinc finger protein ZNF512 考点17:SUV39H家族蛋白 应译为 SUV39H family proteins 考点18:FXN基因 应译为 FXN gene 考点19:共济失调症 应译为 ataxia 考点20:紫杉醇生物合成 应译为 Paclitaxel Biosynthesis 考点21:紫杉醇关键药效基团 应译为 key pharmacophore of paclitaxel 考点22:紫杉醇生物制造 应译为 paclitaxel biomanufacturing 考点23:多年生作物 应译为 perennial crops 考点24:MADS-box基因 应译为 MADS-box genes 考点25:植物寿命长短的关键基因 应译为 key genes determining plant lifespan
7b99c
新闻资讯
新闻报道
80