This article is rated C-class on Wikipedia's
content assessment scale. It is of interest to the following WikiProjects: | ||||||||||||||||||||||||
|
The contents of the DS1C page were merged into Digital Signal 1 on February 26, 2013. For the contribution history and old versions of the redirected page, please see its history; for the discussion at that location, see its talk page. |
I'm a first time editor here, so I hope I'm going about this in the right way. I'm a jack of all trades but a master of only 1, T1 that is. As one of the original 5 T1 carrier sales engineers beginning in 1966, there are many things I can clear up that are anything but clear as we view things today. T1 is truly an amazing resource, less appreciated than it should be, and still capable of achieving more than any other available means of serving the public hunger for integrated services by digital network means. Why ISDN has been so poorly promoted is the primary reason why it hasn't been more popular with the public.
In the early days of T1 deployment it represented an enormous leap forward to applications engineers, cutting the inplace cost of comparable analog systems by a factor of 3, a gap that preceded what we are accustomed to with silicone based modular systems technologies like PCs and home media equipment that gets better and cheaper with each new model that comes along. Meanwhile, analog systems were less well suited to large scale integration of functionality, and by comparison their manufacture escalated costs while the T1 alternative products got cheaper and better day by day.
There were also very significant operational benefits to T1 so long as it was properly deployed, and it took a few years for expertise to catch up with feasibility. But it didn't take long to learn that T1 was worth the effort. By 1970 T1 was among the hottest commodities being sold in the telecom hardware biz, and the rapid evolution of technology and newly learned techniques accelerated acceptance, but leapfrogging became a problem that lingers to this day....
Designers foresaw the need for an aggregation hierarchy which theoretically became known as T2, T3, T4/5, etc. As T1 became the popular choice of technologies the plan was to design and build new cables that would be installed to support T2, a signal that presented a much more difficult transmission challange than the 772 kilohertz wave envelope emitted by T1. Before these cable designs could be mastered the engineering labs were already making noise about optical fiber, so the future seemed headed in that direction and away from copper twisted wire. Interim experimentation with coax and broad guage twisted pairs was conducted during the '70s. The original T2 thru T5 hierarchy was expected to be as follows:
T2=T1x2= 48channels=3.152MBS= 1.5Megahertz T3=T2x2= 96channels=6.312MBS= 3.2Megahertz T4=T3x2=192channels= 13MBS= 6.5Megahertz T5=T4x2=384channels= 26MBS= 13Megahertz
Back when these projections were being contemplated the highest bundling of 4Khz VF channelized multiplex bundles were conveyed either on coaxial tubes or microwave basebands, with the largest bundles known to man at the time being were Master Channel Groups of 300 channels each. As pragmatists began to attempt to configure twisted wire cables that could support the first aggregating step, T2, they struggled with the regenerator capability to pass analog signal envelopes of 3MHz or better through repeaters spaced far enough apart to make be operationally and economically feasible. Meanwhile GTE's transmission manufacturer Lenkurt Electric produced an optical T2 array for electric power companies that operated on fiber optic cable, and Collins Radio installed beta tests of a comparable system using 2 parallel coaxial cables of the kind being used by the fledgling cable TV industry.
Frustration offset success and the industry at large began to recognize that the digital future was optical, and that it was quite a ways off. In the interim T1 popularity continued to grow for a broadening set of reasons, not the least of which was the elimination of frequency sensitive channelizing elements. This in turn allowed channel differentiation to be reassigned to a broadening array of signal interface types, which meant that a T1 channel bank could be configured to host many types of circuit applications on a plug and play basis. The economic and operational impact of this was enormous, but is seldom mentioned by writers of T1 history.
The frame packaging of the 1.544MB 24 CH T1 bundle underwent refinements during this evolutionary period, and channel bank packaging nomenclature known as D types became intermingled with framing definitions. There has been considerable discussion of D1, D1A, D1B, etc., but little mention of the fact that channel bank configuration changes were also part of the ID meaning. By the time we got to D3/D4 we understood SuperFrame bitstream references but much confusion remains about the fact that D4 also implies twin channel banks, or 2xT1 bundles with latest bit frame organization. Not many people realize that ESF is D5 framing. These subtleties only matter if you really want to know why T1 history can be so confusing to people who didn't actually deal with it daily.
My best attempt to curtail this longwinded dissertation will be done here. T1 is arguably the most signficant network transmission protocol in the development of the internet, not to say that TCP/IP is less important, but to simply suggest honest assessment of how digital media have actually evolved. It is also important to recognize that T1/E1 continue to hold great future potential for those who want to delve into how to take a couple of megabits of data transport capacity and find the most practical ways to do positive things by available means. Thanks for your attention; there's a lot more detail in this history if people are interested.
I just have a minor issue about the "Trivia" Section. Yes, T1 and DS1 are quite often used in an interchanble manner. However; T1 actually identifies the physical interface while DS1 is in reference to singnaling. It may be common, but to be blunt; it is more telco slang than anything else to use them in an interchangable manner.
Sorry, don't know who posted the orginal commented here, it wasn't signed.
I see your point, but many companies (Avaya being one of them) use DS1 as their interface indication, it's not so much slang as the context in which it is used.
What I do have a problem with are the pictures in this article, they are very confusing to non telco people, please have a better description of them within the article or find some better easier to understand pictures. Also, would someone mind adding an alarm section or reference an article to the various alarms (red, yellow, blue).
Thanks!
63.87.170.72 17:28, 7 February 2006 (UTC)
How does the physical medium look like? I think its good to have a distict photo of the cross section and terminator jacks, male and female sides. Over what max distance the signal can be tranmitted through such cables?
(response) I agree it would be good to include the physical medium information, but maybe under the "physical medium" link. The answer is pretty complex, since T1 as the end user knows it is probably most often an RJ-45 jack and ALBO (Automatic Line Build Out) levels. Within communication sites it is more often DSX-1 style and level. There is also a lot of DB-15 DSX level interface today.
The RJ-45 interface really should be identified in the same location as the physical medium for 56 Kbps DDS and (n)base-T, since the pinouts are coordinated with those services. Does that sound right?
Flagmichael 01:49, 6 March 2007 (UTC)
22-Aug-2006: This important, technical article (about common Internet technology DS1) was quickly reworded in formal tone by removing "you" and replacing prepositions them/they with nouns, replacing "where as" with "whereas" and using formal grammar and punctuation. The Notes section was added to cite sources in "<ref\>" tag footnotes, per Wikipedia:Guide_to_layout (The "References" section is more tedious, requiring alphabetic by authors). The alternate abbreviation "T-1" was changed to match "T1" wherever used. The diagram errors in Figure 2 ("DS1 SF Frame") were image-edited/uploaded to correct final Terminal bits as "01" + new caption. Beware of other grammar or punctuation problems still in this complex article. - Wikid77 18:33, 22 August 2006 (UTC)
I think it would be good to add a sentence to the SF Bit Robbing paragraph noting that since CAS is frame specific, cross-connections on a digital level with equipment known as a Digital Access and cross-connect system (which see) will lose the signaling bits unless properly handled. Flagmichael 02:00, 6 March 2007 (UTC)
I added this content today. As I am new to Wikipedia editing, feel free to fix it up as necessary!
Flagmichael 18:28, 25 March 2007 (UTC)
Anybody care to start an article on that topic? Jim.henderson 03:20, 22 September 2007 (UTC)
DS1 vs DSL Answer part 1 of 4 Ignoring anything but the technology, DSL seems to get higher bandwidth per line than DS1, by a very large factor; yet I still see people talking about and using DS1. What's the tradeoff? Range? Anaholic ( talk) 15:20, 19 February 2008 (UTC)
A T-span is a portion of the T1 physical wiring between two central offices. Multiple spans are connected together to complete the end to end connection. Repeaters are required between two spans. Spans are typically built to support 25 T1 lines due to cable and repeater case design. The term span can also refer to the full complement of the 25 individual spans. The spans may or may not be "equipped". That is to say the cabling may be in place and connected to the repeater cases with the repeaters installed or the repeater cases may be (partially) empty. Obviously, an individual complete (end to end) span must be equipped to support a working T1. These concepts are not well known as they are only relevant to T1 service providers. Also, spans are becoming less common because T1s are typically provisioned long distance on higher level fiber based systems such as SONET or as a psuedo-wire service on an IP network. Bellhead ( talk) 17:57, 2 January 2009 (UTC)
I'm also a first time editor/contributor. There appears to be several oversights in figure 2, which explains the DS-1 framing format.
Figure 2 seems to have reversed the identification of the terminal framing and signal framing bits.
The odd frames in figure 2 (pink) appear to be mis-identified as the signal framing bits (should be terminal framing).
The even frames in figure 2 (tan/orange) appear to be mis-identified as the terminal framing bits (should be signal framing).
Also, the bit pattern used in figure 2 for the even frames appears to be incorrect. The signal framing bit pattern in the text and in other sources is "001110", ie, the signal frame is identified by the transition from 0 to 1 and from 1 to 0.
Although this seems to be a pretty clear oversight, as a first time editor, I am reluctant to make any changes without hearing feedback from others. (I would also need advice on how to edit the .png image. If I understand correctly, .png images are bit maps, which would make editing more difficult (at least for me) than if the image were available in some other format.)
Lifsorg61 ( talk) 22:02, 31 March 2009 (UTC)
Figure 3 (ESF) very misleading? Shouldn't the Framing bits really be in column 1 (i.e. the left) and the data bits in 2-193? That way the CRC would be computed over the frame diagram from the first Framing bit (row 1, col 1)to the last data bit (row 24, col 193). As it is now the CRC needs to be computed from the row 1, col 193 to row "25", col 192. Having framing in column 1 would also make it consistent with G.704 sect. 2.1.3.1. (I think it would also get the robbed bits in the right place as well.) —Preceding unsigned comment added by 192.31.192.5 ( talk) 19:22, 11 April 2010 (UTC)
The PRI versus inband T1 better/worse argument is a mute point. PRIs weren’t designed to be “faster” than inband T1s. (In fact, for a single T1, it is arguably not true based largely on the average call holding time – which can vary widely depending on application – used in calculations. Do the math.) They were designed to solve other issues.
PRIs are a customer, voice service, bulk delivery mechanism. They are not used internally by telephone companies. One of the original concerns transporting call setup and other signaling information internally (with inband T1s) was speed, but there were many others, notably security. These issues were addressed internally with SS7. Within the telephone company SS7 supports T1s that are neither PRI nor inband T1s. There are various types of SS7 links, (very) loosely comparable to the PRI D channel that carry the signaling information related to the voice channels of these T1s. This was a very effective speed and security solution for the phone company. However, it is inefficient, financially, for a customer to buy SS7 links. A PRI places the voice channels and signaling information in a nice package that is designed to connect directly to the customer’s PBX.
The details supporting the above information are outside the encyclopedic scope appropriate for this article. Read the SS7 and PRI articles for some additional insight.
Bellhead ( talk) 15:51, 21 April 2009 (UTC)
Thanks for taking the time to respond. Your explanations are much more precise than mine as I was trying to generalize for a less technical audience. I agree with most of your input but have provided some specific comments above. Bellhead ( talk) 15:57, 22 April 2009 (UTC)
in the "alternatives" section there is a long commercial for wimax... that shoudl be deleted! —Preceding unsigned comment added by 160.39.129.187 ( talk) 08:45, 31 July 2009 (UTC)
A T1 is a bipolar signal, which is at approx 772KHz This is the signal that will span distances, to the customer for example.
A DS1 is a unipolar signal, which is at approx 1.544 MHz, this signal suffers more attenuation and usually is internal between equipment. —Preceding unsigned comment added by 24.244.73.150 ( talk) 05:21, 26 November 2009 (UTC)
The T designation stands for Terrestrial carrier. Earth bound, not satellite to ground.
since voice is a 4KHz roughly, and sampled at twice thanks to Nyquist the 8KHz is nearly 125 micro seconds. When trying to do TDM when T1 was invented the only method at the time was tubes. A tube can switch roughly at 5.2 microseconds - 125 / 5.2 = 24 voice channels. Europe waited for the transistor and thus the binary 2^5=32 —Preceding unsigned comment added by 24.244.73.150 ( talk) 05:27, 26 November 2009 (UTC)
There is a difference between when you used it-as a finished product-and when a T1 was invented. —Preceding unsigned comment added by 24.244.73.150 ( talk) 23:41, 28 November 2009 (UTC)
well... How did they come up with 24 ? This isn't a power of 2. The first prototype, as I was told, was designed by bell labs with tubes as they where more readily available at that time. As the transistor improved the choice to use transistors in production was no longer a reliability or a cost issue. I could be wrong, but I have yet to hear any other answer how the number 24 was devised other than how fast a tube could switch in the prototype. —Preceding unsigned comment added by 174.6.226.199 ( talk) 01:35, 3 December 2009 (UTC)
Sounds like a typical case of "the Law of the handicap of a head start" to me. Mahjongg ( talk) 00:32, 4 December 2009 (UTC)
I tagged this article as too technical. The reason I did so is because I was looking for information of the internet connection. T1, naturally, goes to a disambig page, so I guessed that this was it. Unfortunately, the introduction to the article was so technical, that I still am not sure if it's the article I'm looking for. Rather than the intro focusing on how it works, it should focus on what it does (that is, provide broadband connections, or phone lines, or something). That would help immensely. D O N D E groovily Talk to me 00:31, 16 January 2012 (UTC)
I was a Member of Technical Staff at Bell Telephone Laboratories, Holmdel, NJ, and part of the T1 Transmission Systems Engineering Group from 1976 to 1984. There are a number of acronyms and definitions being confused with each other on the main Article page. I can imagine the potential for confusion. Will be willing to help a little here and there. [User:tballister]
Hi. Your comments about the technicality of the DS-1 article have really challenged me. I've been studying the organization of this page and its related topics and would really like to help. I spent a number of years engineering T1 at Bell Telephone Laboritories, was published in the Bell System Technical Journal along with some of my colleauges, and I think you're right; this page needs help. There is a signficant amount of mis-representation as well as disorganization.
I've arrived now at several organizational changes that I think will be appropriate and helpful. But the biggest challenge is the one you've put forward, mainly to make this readable to the non-technical user. On one hand I'd argue that it isn't too technical because this page is organized within the "Information Technology Portal", and by defintion technical content generally targets a technical audience. Also, I'd argue that it isn't necessarily the goal of any particular technical discussion to elaborate on the full spectrum of technological premises and foundational concepts required for full understanding. I'd think that if every article did that we'd all be re-writting the same background content over and over again. I mean, I wouldn't think it efficient for everyone who wanted to explain an algebraic operation to always start out with the rules of addition, subtraction, multiplication and division. Right?
On the other hand, as a technical writer for many years, I absolutely recognize numerous opportunites for setting the foundation in this article, generously sprinkled with backward references to other articles that explain the foundational concepts. As as few examples, why does the Digital Signal 1 even exist? Where does it fit within the family of systems that make up the Digital Signal Heirarchy? Why does the heirarchy itslef exist? Who invented it? What problem did it solve?
Guess what I'm saying is that I agree a basic stage has to be set, and this article doesn't do that well, nor do most of the related topics. So I'd like to try and help. But while my opinion is that the best way to do that is indeed, as you suggest, begin within non-technincal discussion, I feel that technical articles must by definition eventually get to the technical points. That means it must at some point begin to rely on technical terminology. But the best articles I've ever read myself are ones that start out with the fundamental concepts in as simple language as possible, and then naturally progress in terms of technological depth, layer by layer. I.e., an approach that insures every reader understands the concepts from the beginning, but then leaves it up to the individual reader's background to let them decide how deeply they want to go.
My concern is that you won't agree with that. Would you let me know your thoughts? I would like to share my organizational recommendations and see if you concur. t Tballister ( talk) 19:14, 20 January 2013 (UTC)