????

Your IP : 3.135.223.175


Current Path : C:/Program Files/Microsoft SQL Server/150/DTS/Binn/Resources/1033/
Upload File :
Current File : C:/Program Files/Microsoft SQL Server/150/DTS/Binn/Resources/1033/dtspipeline.rll

MZ����@���	�!�L�!This program cannot be run in DOS mode.

$�����ծ��ծ��ծ���Q��ծ���9��ծ������ծ�Rich�ծ�PEd�
��]�" �@�]d`A ��h>.rdatap@@.rsrc� � @@
��]
TGCTL.rdataT.rdata$zzzdbg @.rsrc$01@7��.rsrc$02 �@�p����0�H�e`��x�-��.������o���
 �
8�P�h�������?��@��E��F�G(�H@�rX��p������������������30�4H�r`�sx����������4��l��m�n ��8��P��h����'��(��f���������	�!(	�"@	�`X	�ap	�q�	�r�	�s�	���	���	��
�
�0
�"H
�Z`
��x
���
���
���
���
���
���� ��8��P��h���T��U��������������(�	@�	X�	p�N	��T	���	���	��	
��


�H

�O
0
��
H
��
`
��
x
��
��
��
��
��
�B�H �I8��P��h����������������������(��@�jX�p���	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�			 	0	@	P	`	p	�	�	�	�	�	�	�	�		@7�>��@E��G�PXh�]�dp�gPc�pd`�eHf�����i��n��z�`{H�~V�v�rx��X�x�:��F������ ���(�Px�|���ز@���� �\�
 � @�� �Hh�N���h�d��
�������Dh���P8�P��@�h��	@�$�����`��P���L�D�~�8�X0b��p~�2p�(B
P)�P �P.�0|�1�3h�'bp� 4Xx;�=��?��F2�H��P
������Z��]�^��a,�f��it8j
0��F�����8�8H�������Ȟ:����x��(�@�$Hr��x������h�������>����\��\"���h�}REGISTRYHKCR
{
   SSIS.Pipeline.7 = s 'MainPipe Object'
   {
      CLSID = s '{D9DA2996-AD06-4166-9C0C-C53E29834AF2}'
   }
   SSIS.Pipeline = s 'MainPipe Object'
   {
      CLSID = s '{D9DA2996-AD06-4166-9C0C-C53E29834AF2}'
      CurVer = s 'SSIS.Pipeline.7'
   }
   NoRemove CLSID
   {
      ForceRemove {D9DA2996-AD06-4166-9C0C-C53E29834AF2} = s 'Data Flow Task'
      {
         ProgID = s 'SSIS.Pipeline.7'
         VersionIndependentProgID = s 'SSIS.Pipeline'
         ForceRemove 'Programmable'
         InprocServer32 = s '%MODULE%'
         {
            val ThreadingModel = s 'free'
         }
         'TypeLib' = s '{36386939-825E-4817-ADEF-9FD04249EE91}'
         ForceRemove 'Implemented Categories'
         {
            ForceRemove {30DEA136-5D17-4352-B6CB-E2D8931B1B6C}
         }   
         ForceRemove DTSInfo
         {
             val Description = s 'Performs high-performance data extraction, transformation and loading'
             val TaskType = s 'DTS100'
             val TaskContact = s 'Performs high-performance data extraction, transformation and loading;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights Reserved;http://www.microsoft.com/sql/support/default.asp;1'
             val UITypeName = s 'Microsoft.DataTransformationServices.Design.PipelineTaskView, Microsoft.DataTransformationServices.Design, Version=%MANAGEDVERSION%, Culture=neutral, PublicKeyToken=89845dcd8080cc91'
             val ResourceFile = s 'DTSPipeline,108,109'
             val HelpKeyword = s 'sql13.dts.designer.dataflowtask.F1'
             val SamplesTag = s 'SsisDataflowTask'
         }
         DefaultIcon = s '%MODULE%,0'
      }
   }
}
	
HKCR
{
   DTS.BufferManager.7 = s 'CDTSBufferManager Object'
   {
      CLSID = s '{383E62BE-EC63-4D9C-A0E0-B89E2618AAEF}'
   }
   DTS.BufferManager = s 'CDTSBufferManager Object'
   {
      CLSID = s '{383E62BE-EC63-4D9C-A0E0-B89E2618AAEF}'
      CurVer = s 'DTS.BufferManager.7'
   }
   NoRemove CLSID
   {
      ForceRemove {383E62BE-EC63-4D9C-A0E0-B89E2618AAEF} = s 'CDTSBufferManager Object'
      {
         ProgID = s 'DTS.BufferManager.7'
         VersionIndependentProgID = s 'DTS.BufferManager'
         ForceRemove 'Programmable'
         InprocServer32 = s '%MODULE%'
         {
            val ThreadingModel = s 'free'
         }
         'TypeLib' = s '{36386939-825E-4817-ADEF-9FD04249EE91}'
      }
   }
}
	
HKCR
{
   DTS.ManagedComponentWrapper.7 = s 'Managed Component Host'
   {
      CLSID = s '{7CDF593F-DE06-4ABD-B356-7976EF7AC8E0}'
   }
   DTS.ManagedComponentWrapper = s 'Managed Component Host'
   {
      CLSID = s '{7CDF593F-DE06-4ABD-B356-7976EF7AC8E0}'
      CurVer = s 'DTS.ManagedComponentWrapper.7'
   }
   NoRemove CLSID
   {
      ForceRemove {7CDF593F-DE06-4ABD-B356-7976EF7AC8E0} = s 'Managed Component Host'
      {
         ProgID = s 'DTS.ManagedComponentWrapper.7'
         VersionIndependentProgID = s 'DTS.ManagedComponentWrapper'
         ForceRemove 'Programmable'
         InprocServer32 = s '%MODULE%'
         {
            val ThreadingModel = s 'free'
         }
         'TypeLib' = s '{36386939-825E-4817-ADEF-9FD04249EE91}'
         ForceRemove 'DTSInfo'
         {
            val ComponentType = d 0
			val CurrentVersion = d 0
            ForceRemove 'EN-US' 
            {
               val Description = s 'Managed Component Host'
            }
         }
      }
   }
}
	
HKCR
{
   DTS.ErrorComponent.7 = s 'Error Component'
   {
      CLSID = s '{FDA22D13-BECE-483E-88B7-64684DEE9CD7}'
   }
   DTS.ErrorComponent = s 'Error Component'
   {
      CLSID = s '{FDA22D13-BECE-483E-88B7-64684DEE9CD7}'
      CurVer = s 'DTS.ErrorComponent.7'
   }
   NoRemove CLSID
   {
      ForceRemove {FDA22D13-BECE-483E-88B7-64684DEE9CD7} = s 'Error Component'
      {
         ProgID = s 'DTS.ErrorComponent.7'
         VersionIndependentProgID = s 'DTS.ErrorComponent'
         ForceRemove 'Programmable'
         InprocServer32 = s '%MODULE%'
         {
            val ThreadingModel = s 'free'
         }
      }
   }
}
	
( @ @j�>h��g���Y���Gp��:\y�:\y�:\y�:\y�/Sr�.Rq@j�>h��g���Y���Gp��:\y�:\y�:\y�:\y�/Sr�.Rqf���X��������������u��Y��;}��:\y�f���X��������������u��Y��;}��:\y�f���r������������u��b��L��<_|�f���r������������u��b��L��<_|�g���r������������u��b��L��?c��<wF�f���r������������u��b��L��?c��i���r������������u��b��L��Bi��<wF�<wF�f���r������������u��b��L��Bi��j���r������������u�������������<wF�<wF�<wF�����r������������u��b��L��Fn��l���r������������u�����E�V�E�V�E�V�E�V�C�R�A�O�@�M�>|J�=zH�<wF�<wF�<wF����������������u��b��L��Jt��n���r������������u�����J�_�J�_�J�_�J�_�H�[�F�X�E�T�B�Q�A�N�?~K�>{H�<xG�<wF��������������u��b��L��O{��q���r������������u�����N�f�N�f�N�f�N�f�M�c�K�`�I�]�G�Y�E�V�C�R�A�O�@�M�>|J�=yH������������u��b��L��S���s���r������������u�����R�m�R�m�R�m�R�m�P�k�P�h�M�e�L�b�J�^�H�[�F�X�D�T�C�Q��������������u��b��L��W���u���r������������u�����S�n�S�n�S�n�S�n�S�n�S�m�Q�k�P�i�O�g�M�d�J�`�I�]����������������u��b��L��[���x���r������������u�������������R�m�Q�j�O�h�����r������������u��b��L��^���z��r������������u��b��L��b���S�n�S�n�x���r������������u��b��L��b���{��r������������~��b��L��e���S�n�{��r������������~��b��L��e���}�����������������}��g������}�����������������}��g�����������������������g�����������������������g���������������������t��g���������������������t��g���z��z��p���������{��v��q���n���e��pb��z��z��p���������{��v��q���n���e��pb������������������������������������������������������x����������������������������������(  ����J~��R���L���@��,{������m�v�����J~��R���L���@��,{������6|��R������������m�v� �����z��B��
��(q��7���N�������q�z�) �&2(�.=0�5F7�F�W� ����y��C����+o��;���M������}������ڞ�{ѕ�qƋ�g��T�h� ����C����+u��;���M����������������۬��ש�g��K�[����D����+u��<���H�������������}���w���q�z��ܮ�o�y����p��5����-v��?���^�������������w���o�y��������x��F��-z��m�����������������_�������������������������_���a��v�������������e���a��v�������������e���Y��]��qf���g���f���]���Y��]��qf���g���f���]�����������������������y��������   � h
%ls Input %lu%ls Output %luInput column lineage IDVariable used.Output column lineage ID	ComponentDTSPipelineManaged Component Wrapper?Wrapper component for loading managed SSIS data flow components\This property specifies Assembly Qualified Name of user's managed data flow component class.�This property specifies the name of the Visual Studio for Applications project when a managed component wrapper (MCW) is used to host the script transformations.Connection usedData Flow TaskMoves data between sources and destinations while transforming and cleaning. Data can be moved between tables and files while efficently processing the data in memory with transformations. This is an essential Integration Services task and may be the only task in the control flow.Prepare for ExecutePre-Executeoutputerrordestination adaptercolumn	componentcomponent viewerpathinputPost ExecuteCleanup
ValidatingInputOutputinput columnsource adapter
output columnpropertyruntime connection	transformunspecifiedunknown object type (%1!d!)
virtual inputvirtual input columnexternal column	ErrorCodeErrorColumn&Path "%1!s!" (%2!d!) has been deleted.ALayout changes might have made component "%1!s!" (%2!d!) invalid.
%1!s! [%2!d!].Data flow engine will call a component method.;Data flow engine has finished a call to a component method.%1!s! "%2!s!" (%3!d!)"%1!s!" (%2!d!)
.%1!s![%2!s!]Paths[%1!s!.%2!s!]InputsOutputsColumnsExternalColumnsConnections
Properties'The default buffer size is %1!d! bytes.(Buffers will have %1!d! rows by default.wDirectory "%1!s!" could not be used for temporary storage. Error %1!d! was encountered while writing to that directory.xNo temporary buffer storage locations are available. The data flow will fail if not enough physical memory is available.{No temporary BLOB data storage locations are available. The data flow will fail if not enough physical memory is available.�No temporary buffer storage locations were provided. The buffer manager will consider the directories in the TEMP and TMP environment variables.�No temporary BLOB data storage locations were provided. The buffer manager will consider the directories in the TEMP and TMP environment variables.hThe data flow will not remove unused components because its RunInOptimizedMode property is set to false.�Rows in buffer type %1!d! are larger than the configured maximum for the entire buffer size. There will be only one row in buffers of this type.�Rows in buffer type %1!d! would cause a buffer size greater than the configured maximum. There will be only %2!d! rows in buffers of this type.�Rows in buffer type %1!d! would cause a buffer size less than allocation minimum, which is %2!d! bytes. There will be %3!d! rows in buffers of this type.An unidentified componenta%1!s! leaked a buffer with ID %2!d! of type %3!d! with %4!d! rows and a reference count of %5!d!.TThis is a virtual buffer referencing the physical buffer with ID %d for its storage."There are no trees to be executed.The execution plan is empty.?Automatically adjust buffer size to match max rows in a buffer.NThe buffer size of buffer type %1!d! is automatically adjusted to %2!d! bytes.MThe %1!s! spent %2!llu! milliseconds in PrimeOutput filling buffers on %3!s!.|During last execution the pipeline suspended %2!s! of %1!s! for %3!u! milliseconds to limit the number of in-memory buffers.TBuffer manager is throttling allocations to keep in-memory buffers around %1!lu!MB

VMemory pressure was alleviated, buffer manager is not throttling allocations anymore

�Failure while loading back a swapped bufer. It may mean the system is in a low memory state and the buffer cannot get its memory allocated.Begin Path %1!lu!

Begin Subpath %1!lu!

End Path %1!lu!



End Subpath %1!lu!

Unknown object

; Unknown component

,Begin output plan

   Begin transform plan

/   End transform plan



   Begin source plan

'   End source plan

End output plan



Begin path plan

End path plan

Begin Path Plan %1!lu!

Begin Subpath Plan %1!lu!

End Path Plan %1!lu!



End Subpath Plan %1!lu!

Call ProcessInput oncomponent (%1!lu!) for input (%1!lu!)

.Create new execution item for subpath %1!lu!

Create new row view for %1!s!

Call PrimeOutput on output (%1!lu!)

 unknown output

2Various information from Data Flow initialization.After execution, some buffers were left outstanding. Some transform is hanging onto buffer references and if it freed them we would use a lot less memory. The logged text includes information about the transform that owned the buffer, and the ID of the buffer.{The Data Flow changed the size of a buffer away from the default values. The event text includes the sizes and the reasons.<The input the scheduler had when forming the execution plan.$The execution plan of the data flow.�Information about each phase of validation and execution for each data flow component. The processing time of each phase is specified./The %1!s! spent %2!I64d! milliseconds in %3!s!.3A component has returned from its PrimeOutput call.*PrimeOutput will be called on a component.5Rows were provided to a data flow component as input.3A component will be given the end of rowset signal.4A component has finished processing all of its rows..Numeric identifier of the data flow component. Name of the data flow component.2The buffer object that was given to the component.)Numeric identifier of the data flow path.Name of the data flow path..Numeric identifier of the data flow component. Name of the data flow component. Numeric identifier of the input.Name of the input..Numeric identifier of the data flow component. Name of the data flow component. Numeric identifier of the input.Name of the input..Numeric identifier of the data flow component. Name of the data flow component.Name of the method.!Start time of the component call.End time of the component call.�%1!s!;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights Reserved; http://www.microsoft.com/sql/support;%2!ld!YExternalRequest_pre: The object is ready to make the following external request: '%1!s!'.BExternalRequest_post: '%1!s!'. The external request has completed.=Specifies the ID of the input column used in the aggregation."Specifies the type of aggregation.�Specifies the approximate number of group by keys that the aggregation produces. Low indicates 500,000 key values, Medium indicates 5 million key values, and High indicates more than 25 million key values. Unspecified indicates no KeyScale value is used.�Specifies the exact number of group by keys that the aggregation produces. If a KeyScale value is specified, the value in Keys takes precedence.tSpecifies the percentage by which memory can be extended during the aggregation. Valid values are between 1 and 100.�Aggregates data with functions such as Average, Sum, Count, Max, and Min. For example, group data to compute the sum of sales per product or the average visits per day.Aggregate Input 1Aggregate Output �Indicates whether the column requires special handling because it may contain oversized data or it requires precision beyond a float data type.(Specifies the string comparison options.NIndicates whether a warning is posted when an arithmetic overflow is detected.	AggregateGroup byCount	Count allCount distinctSumAverageMinimumMaximumUnspecifiedLowMediumHighSpecifies the approximate number of distinct values in a column that the aggregation counts distinctly. Low indicates 500,000 key values, Medium indicates 5 million key values, and High indicates more than 25 million key values. Unspecified indicates no CountDistinctScale value is used.�Specifies the exact number of distinct values in a column that the aggregation counts distinctly. If a CountDistinctScale value is specified, the value in CountDistinctKeys takes precedence.Joins additional columns to the data flow by looking up values in a table. For example, join to the 'employee id' column the employees table to get 'hire date' and 'employee name'. We recommend this transformation when the lookup table can fit into memory.Type of Reference Data Source.9[InputColumnName,ReferenceColumnName]; one or more times.A[ReferenceColumnName,InputOrOutputColumnName]; one or more times.<Specifies the SQL statement that generates the lookup table.LSpecifies a SQL statement that uses parameters to generate the lookup table.Runtime Connection Name..Connection manager used to access lookup data.-Specifies the cache type of the lookup table.LookupFullPartialNone>Maximum Memory Usage for Reference Cache on a 64 bit platform.FSpecifies the type of connection used to access the reference dataset.OLEDBCachehSpecifies how the Lookup transformation handles rows without matching entries in the reference data set..Treat rows with no matching entries as errors.:Send rows with no matching entries to the no match output.Lookup No Match Output�The Lookup output that handles rows with no matching entries in the reference dataset. Use this output when the NoMatchBehavior property is set to "Send rows with no matching entries to the no match output."sSpecifies the percentage of the cache that is allocated for rows with no matching entries in the reference dataset.qDetermines whether duplicate keys in the reference data should be treated as errors when full cache mode is used.>Maximum Memory Usage for Reference Cache on a 32 bit platform.9Indicates whether to reference metadata in an XML format.BConvert joined and copied column datatypes to match automatically.�Specifies the list of lineage identifiers that map to the parameters that the SQL statement in the SQLCommand property uses. Entries in the list are separated by semicolons.Lookup InputLookup Match OutputLookup Error OutputJSpecifies the column in the reference table from which a column is copied.@Specifies the column in the reference table that a column joins.`Set to True if Lookup should validate reference metadata during validate, false if it should notCopy ColumnLookup�Copies input columns to new columns in the transformation output. For example, create a copy of the 'address' column so that it can be cleaned by subsequent transformations and leave the original column in place for auditing.7Specifies the lineage identifier of the column to copy.Copy Column InputCopy Column OutputCopy Column�Applies string functions to character data. For example, convert from data that is in upper case to lower case. Operates only on column data with a string data type.Character Map InputCharacter Map Output)Specifies the string operations to apply.OSpecifies the lineage identifier of the input column to which operations apply.Character Map Error Output
Map Column
Character Map	Lowercase	Uppercase
Byte reversalHiraganaKatakana
Half width
Full widthLinguistic casingSimplified ChineseTraditional ChineseExpands an un-normalized data flow into a more normalized version. Values from multiple columns of a single record expand to multiple records in a single column. For example, expand a table that lists athletic scores with a column for each game to more records and a single score column.
Unpivot InputUnpivot OutputUnpivotbIndicates whether the value in the PivotKeyValue property is written to the column in the new row.�The lineage ID of the output column that will receive the value from this input column. Or -1 if column's value is not used in output rows.ESpecifies the value written to the PivotKey column in the output row.Deprecated Union All TransformDeprecated Union All Input %dDeprecated Union All Output;Specifies the maximum number of threads to use for sorting.SortRIndicates whether the Sort transformation removes rows with duplicate sort values.Sort Output
Sort Input4Specifies the comparison options for character data.'Specifies the sort order of the column.`Specifies the lineage identifier of the input column that is associated with this output column.Sorts input data in ascending or descending order when it cannot be sorted at its source. For example, sort sales data by 'customer id' before it is written to a flat file and delivered to another department. Note: We do not recommend this transformation for large data flows.Slowly Changing Dimension3Updates a SQL Server Analysis Services data warehouse dimension. Update dimension records that are already present and  insert dimension records that aren't. For example, update the DimProduct table of the AdventureWorksDW OLAP database with data from the Products table in the AdventureWorks OLTP database.Slowly Changing Dimension Input,Input to Slowly Changing Dimension TransformUnchanged OutputUnchanged rows
New OutputNew rowsFixed Attribute Output2Rows with changes in columns tagged FixedAttribute!Changing Attribute Updates OutputPThese rows contain new information for ChangingAttribute (Type 1) column changes#Historical Attribute Inserts Output=HistoricalAttribute (Type 2) changes result in these new rowsInferred Member Updates Output7Information to update unpopulated initial dimension rowLookupConnection@Defines the connection to use for lookups in the dimension tableAll newDetectOtherKeyChanging attributeHistorical attributeFixed attribute>Specifies the SELECT statement used to create a schema rowset.[Specifies that all rows in the input are new or the transformation detects the change type.xIndicates whether historical attribute updates are directed to the transformation output for changing attribute updates.xSpecifies the WHERE clause in the SELECT statement that selects the current row among rows with identical business keys.^Indicates whether the transformation fails when columns with fixed attributes contain changes.UIndicates whether the transformation fails when a lookup of an existing record fails.7Indicates whether inferred member updates are detected.2Specifies the column name for the inferred member.
Compacts an input data flow by pivoting it on a column value, making it less normalized. The input data should be sorted by the pivot column because a pivot happens each time data in the pivot column changes. Note: duplicate rows will cause this transformation to fail.Pivot Default InputPivot Default OutputPivot Error OutputPivot Key LookupPivotsDescribes how the input column participates in the pivot operation. It can help define a set of related records, it can hold the key that uniquely identifies a row within the related set and upon which the values are pivoted, it can hold a value that gets transferred into a pivoted column or it can contain a value that simply passes straight though the pivot unchanged.�The lineage identifier of an input column that contains a PivotedValue or the value -1. The value -1 indicates the column is not used in a pivot operation�Specifies a value that matches one of the possible values from the input column that is marked as the pivot key by the PivotUsage property.(Can't Convert to String)Not PivotedSet Key	Pivot Key
Pivoted Value(Causes the Pivot Transform to ignore rows containing unrecognized values in the Pivot Key column and to report the complete list of unmatched values at the end of DataFlow execution. Useful for collecting a list of values from input rows that can be used to automatically generate output columns.Merge Join Left InputMerge Join Right InputMerge Join Output
Merge Join&Specifies the type of join to perform.4Specifies the number of columns in the ordering key.>Indicates whether the null values are treated as equal values.Full outer joinLeft outer join
Inner join�Combine two sorted data flows into one using the FULL, LEFT, or INNER join. We recommend this transformation when data flows can be sorted at their sources.3The input column associated with this output columnBIndicates the requested maximum buffers per input before blocking.+Specifies the number of rows in the sample.,Specifies the percentage size of the sample.3Specifies the seed for the random number generator.Sampling Selected OutputSampling Unselected OutputOutput for rows in the sampleRow Sampling�Randomly samples a specific number of rows from an input data flow. For example, use it randomly select products to feature on your Web site.Percentage Sampling�Randomly samples a percentage of rows from an input data flow. For example, use it twice to help create a data mining model: create one sample for training the model and another sample for testing the model.-Specifies the variable to hold the row count.	Row Count�Counts rows as they pass through a data flow. For example, use the number of rows processed to determine whether or not to perform cleanup operations, or update text in an e-mail message to communicate processing status.Multicast Input	Multicast�Distributes every input row to every row in one or more outputs. For example, branch your data flow to make a copy of data so that some values can be masked before sharing with external partners.Multicast Output %%d
Merge Input 1
Merge Input 2Merge OutputMerge
Combines rows from multiple sorted data flows into one sorted data flow. Unlike the Merge Join, all rows flow to the output. For example, combine two sorted lists of customers into a single sorted list. If sorting is not important, use the Union All transformation. FThe lineage ID of the output column associated with this input column.0Specifies which audit item this column produces.�Includes data about the environment in which the package runs. Map system variables to new output columns. For example, add the name of the computer and operator to the data flow.AuditExecution instance GUID
Package IDPackage name
Version IDExecution start timeMachine name	User name	Task nameTask IDKSpecifies the ID of the input column. The value must be a positive integer.Fuzzy Grouping InputFuzzy Grouping OutputFuzzy Grouping�Identifies potential duplicate rows and helps standardize the data by selecting canonical replacements. For example, reduce the occurrences of typos in a 'company name' column by referencing an approved list of company names.QSpecifies the minimum similarity threshold, expressed as a value between 0 and 1.bSpecifies the white-space and punctuation characters used to separate strings into multiple words.LIndicates whether to use the column to determine if two records are similar.ZSpecifies whether rows must be exact matches or fuzzy matches to be considered duplicates.�Specifies the hint that causes similar rows that contain differing leading or trailing numeric values be less likely to be considered duplicates.|Specifies the relative contribution of a match in this column to the similarity score. The value must be a positive integer.�Specifies the minimum textual similarity required between input and fuzzy matched records on values in this particular join column.  Must be a floating point value between 0.0 and 1.0.2Specifies how string data in a column is compared.Specifies the column type. The valid values are Canonical, ColumnSimilarity, KeyIn, KeyOut, PassThru, Similarity and Undefined.NumeralsNotSpecialLeadingNumeralsSignificantTrailingNumeralsSignificant%LeadingAndTrailingNumeralsSignificantReading input recordsFinding similar records(Picking canonical records for each groupWriting output recordsFuzzy Grouping Inner Data Flow�Specifies the maximum amount of memory (MB) that transform should use.  A value of 0 will enable dynamic memory usage based upon requirements and available system memory.&If true, each input record will be directly compared against EVERY other input record.  This may produce slightly better results, but will be extremely slow unless the number of input records is very small.  This option is primarily intended for debugging purposes and should be used with care.FullySensitive
IgnoreCaseIgnoreNonSpaceIgnoreKanaTypeIgnoreWidth
IgnoreSymbolsExactFuzzy	UndefinedKeyInKeyOut
SimilarityColumnSimilarityPassThru	Canonical?Specifies the ID of the output column to receive the file data.0TRUE if the file is expected to begin with a BOMImport Column Error Output
Import Column�Reads data from files and adds it to columns in a data flow. For example, use an input column of file paths to import photographs from a directory.Import Column OutputImport Column Input	File Read	File Open�Reads data from a data flow and insert it into a file. For example, create image files based on a database column that contains photographs.Export Column OutputExport Column InputASpecifies the ID of the input column that contains the file data.AIndicates whether to write a byte-order marker (BOM) to the file.5Indicates whether to append data to an existing file.:Indicates whether to force truncation of an existing file.Export Column Error Output
Export Column
File Write	File OpenDerived Column�Creates new column values by applying expressions to transformation input columns. Create new columns or overwrite existing ones. For example, concatenate the values from the 'first name' and 'last name' column to make a 'full name' column.Derived Column Output3Default Output of the Derived Column TransformationDerived Column Error Output1Error Output of the Derived Column TransformationDerived Column Input*Input to the Derived Column TransformationDerived Column Expression"Derived Column Friendly ExpressionComputationComputationData ConversionData Conversion InputData Conversion OutputData Conversion Error Output
ConversionISpecifies the input column used as the source of data for the conversion.NIndicates whether the column uses the faster, locale-neutral parsing routines.�Converts data from one data type to another. Converted data is written to a new output column. For example, convert data from a column of decimal type to an output column of integer type or convert ANSI string data to Unicode.Union All Input %1!lu!Union All Output	Union All�Combines rows from multiple data flows without sorting. For example, add rows back to a data flow after correction of errors. If sorting is important, we recommend using the Merge transformation.FThe lineage ID of the output column associated with this input column.Fuzzy Lookup�Indicates whether to reuse an existing fuzzy-match index table or build a new one. The new index table can be temporary, persisted, or persisted and maintained.ZSpecifies the name of the reference table or view against which to match incoming records.�Specifies the name of the fuzzy-match index table to be reused or the table name in which a newly generated fuzzy-match index table will be written.�Indicates whether to drop an existing fuzzy-match index table if transformation is configured to generate a new table and a table with the same name already exists.�Whether a copy of the reference table should be made for fuzzy match index construction and subsequent matching.  It is strongly recommended that a copy always be made, otherwise unpredictable output may result if the contents of the reference table change during package execution (or between executions if reusing a fuzzy match index).  If this property is set to false, the reference table must additionally have a clustered index built upon an integer identity column.<Specifies the metadata of the reference table in XML format.LSpecifies the maximum number of fuzzy matches to return for an input record.fSpecifies the minimum similarity score required for the match results that the transformation returns.�Specifies the set of white-space and punctuation characters used to separate fuzzy match fields into the words used for matching and scoring a fuzzy lookup.�Specifies the maximum amount of memory (MB) that this instance of the Fuzzy Lookup transformation can use.  If 0 is given, then the transform will dynamically adjust it's memory usage based upon it's requirements and available system memory.:Performs data cleaning tasks such as standardizing data, correcting data, and providing missing values.  Uses fuzzy matching to return one or more close matches from a reference table. For example, lookup customer data by address, even if your input address data does not exactly match your reference address data.�Microsoft SQL Server connection and database in which the fuzzy lookup reference table is located and where the fuzzy match index table will be stored and accessed.pSpecifies the comparison options for character data. The options are those supported by the FoldString function.ZFlags to FoldString that should be used when comparing whether two strings are equivalent.|Specifies the relative contribution of a match in this column to the similarity score. The value must be a positive integer.�If true, Fuzzy Lookup will partially load the error-tolerant index and the reference table into main memory before execution begins.  This can greatly enhance performance if many input rows are to be processed.�If true, each input record will be directly compared against EVERY record in the reference table.  This may produce slightly better results, but will be extremely slow unless the number of records in the reference table is very small.  Note that the entire reference table will be loaded into main memory if this option is enabled.  This option is primarily intended for debugging purposes and should be used with care.Fuzzy Lookup InputFuzzy Lookup Output(Taking a snapshot of the reference tableBuilding Fuzzy Match IndexWarming cachesRows processed by Fuzzy LookupRows Processed�Specifies the minimum textual similarity required between input and fuzzy matched records on values in this particular join column.  Must be a floating point value between 0.0 and 1.0.JSpecifies whether the column is used to indicate confidence or similarity.QName of the reference table column with which the input column should be matched.9Specifies whether the match is a fuzzy or an exact match.XSpecifies the reference table column that is copied to the output for each match result.mSpecifies the lineage identifier of the input column from which to copy to this output column for each match.LCID of the input string data.ReuseExistingIndexGenerateNewIndexGenerateAndPersistNewIndexGenerateAndMaintainNewIndexFullySensitive
IgnoreCaseIgnoreNonSpaceIgnoreKanaTypeIgnoreWidth
IgnoreSymbols	NoMappingMapPrecomposedMapFoldCZone
MapFoldDigitsMapExpandLigaturesFuzzyExact	Undefined	Undefined
Similarity
ConfidenceCandidateScoreNextCandidateScore	ThresholdFullTokensFetched
QGramsFetched
CandidatesRefTuplesFetched_FullTokens_FullTokensIndexed_FullTokensFetched_QGrams_QGramsIndexed_QGramsFetched_Candidates_RefTuplesFetchedInvalidColumnTypeCandidateWeight SuccessiveUnsuccessfulRefFetches_HighFreqTokensFetched_FFMDuration_NextBestScoreColumnSimilarityFullFieldsFetched_FullFields_FullFieldsIndexed_FullFieldsFetchedTerm Lookup&Determines how frequently specific terms occur in a data flow. Matching terms and their frequency of occurrence are written to output data columns. For example, see if a document contains terms that are common in a particular document category. Note: you must supply a table of reference terms.Term Lookup Input#Input to Term Lookup TransformationTerm Lookup Output!Output from Term Lookup Transform@Specifies the reference table that contains the reference terms.Specifies the column that contains the reference terms. The column must be in the table specified in the property RefTermTable.�Specifies whether the column is a pass-thru column and whether it is a lookup column. 0 indicates a pass-thru column, 1 indicates a lookup column, and 2 indicates the column is both a pass-thru and a lookup column.RSpecifies the lineage ID of the upstream column that is referenced by this column.Term from the reference data.9The frequency of the term appearing in every text string.9Indicates whether the lookup operation is case-sensitive.Term Lookup Error Output'Error Output from Term Lookup TransformTerm LookupTerm ExtractionExtracts frequently used English-only terms from an input data flow. Terms and their corresponding score are written to output data columns. For example, search a group of documents to discover commonly-used terms. Note: an English dictionary is included with Integration Services.Term Extraction Input'Input to Term Extraction TransformationTerm Extraction Output*Output from Term Extraction TransformationTerm,Output Column from Term Extraction TransformReference Data Connection.)Indicates whether to use exclusion terms.=Specifies the reference table containing the inclusion terms.YSpecifies the column containing the inclusion terms. The column should be in InTermTable.=Specifies the reference table containing the exclusion terms.|Specifies the column containing the exclusion terms. The column must be in the table specified in the property OutTermTable.xSpecifies the type of terms to extract. 0 specifies noun, 1 specifies noun phrase, and 2 specifies noun and noun phrase.BSpecifies the frequency threshold above which a term is extracted.0Specifies the maximum number of words in a term.@Specifies whether the Term Extraction process is case sensitive.Term Extraction Error Output+Error Output from Term Extraction TransformTerm ExtractionScore,Output Column from Term Extraction Transform�Specifies what kind of score to be associated with a term. 0 means using the frequency of a term as the score; 1 means using the TFIDF of a term as the score.SQL Server DestinationSQL Server Destination Input#SQL Server Destination Error Output2The name of the database object to copy data into.�Specifies that any constraints on the destination object are checked during the data copy operation. Corresponds to the bulk insert CHECK_CONSTRAINTS option.^Specifies the number of the first row to copy. Corresponds to the bulk insert FIRSTROW option.�Specifies that any insert triggers defined on the destination object will execute during the data copy. Corresponds to the bulk insert FIRE_TRIGGERS option.�Indicates whether the values supplied for identity columns will be copied to the destination. If false, values for identity columns will be auto-generated at the destination. Corresponds to the bulk insert KEEPIDENTITY option.�Indicates whether the columns containing null will have null copied to the destination. If false, columns containing null will have their default values inserted at the destination. Applies only if fast load is turned on.\Specifies the number of the last row to copy. Corresponds to the bulk insert LASTROW option.�Specifies the maximum number of syntax errors and compilation errors that can occur before the data copy operation is canceled. Each row that cannot be copied is ignored and counted as one error. Corresponds to the bulk insert MAXERRORS option.[Specifies how the data to be copied is sorted. Corresponds to the bulk insert ORDER option.�Specifies that a table-level lock is acquired for the duration of the data copy operation. Corresponds to the bulk insert TABLOCK option.�Specifies the file used to collect rows that have formatting errors. These rows are copied into this error file "as is."  Corresponds to the bulk insert errorfile option.�Specifies in seconds the amount of time the SQL Destination adapter will wait before terminating if there is no data available for insertion.  A timeout value of 0 means the SQL Destination adapter will not time out.�Provides connections to a local SQL Server database and bulk-loads data into SQL Server tables and views. Note: to optimize performance, we recommend using the OLE DB destination instead.�Specifies when commits are issued during data insertion.  A value of 0 specifies that one commit will be issued at the end of data insertion.3Specifies the variable that contains the recordset.�Creates and populates an in-memory ADO recordset that is available outside of the data flow. Scripts and other package elements can use the recordset. For example, use a recordset to store the names of files that will be loaded into the data warehouse.Recordset Destination InputRecordset Destination	File nameFile name from variable+Specifies the mode used to access the file.0Name of the variable that contains the file nameReads raw data from a flat file that was previously written by the Raw File destination. For example, allow packages to run faster by reading from a local raw file instead of a slow source or manually distribute data for parallel processing on multiple computers.!Specifies the name of the source.Output from %sRaw File Source OutputRaw File SourceOutput file name3Indicates what option is used to write to the file.�Writes raw data that will not require parsing or translation. For example, allow packages to run faster by reading from a local raw file instead of a slow source or manually distribute data for parallel processing on multiple computers.Raw File DestinationRaw File Destination InputTFile %1!s! already exists. The raw file destination adapter will append to the file.SFile %1!s! already exists. The raw file destination adapter will truncate the file.QFile %1!s! does not exist. The raw file destination adapter will create the file.
Create AlwaysCreate OnceAppendTruncate and Append
OLE DB SourceOLE DB Source OutputOLE DB Source Error Output
ConversionEThe mappings between the parameters in the SQL command and variables.�Extracts data from an OLE DB-compliant relational database. Extract from a database table or view, or use a SQL command. For example, extract data from tables in Microsoft Office Access or SQL Server databases.OLE DB DestinationOLE DB Destination InputOLE DB Destination Error OutputInsert�Indicates whether the values supplied for identity columns will be copied to the destination. If false, values for identity columns will be auto-generated at the destination. Applies only if fast load is turned on.�Indicates whether the columns containing null will have null inserted in the destination. If false, columns containing null will have their default values inserted at the destination. Applies only if fast load is turned on.USpecifies options to be used with fast load.  Applies only if fast load is turned on.�Specifies when commits are issued during data insertion.  A value of 0 specifies that one commit will be issued at the end of data insertion.  Applies only if fast load is turned on.Loads data into an OLE DB-compliant relational database, such as SQL Server. Many types of databases are OLE DB-compliant. With minor reconfiguration, this destination can be used before and after your database is migrated from one OLE DB-compliant technology to another. OLE DB Command Error OutputOLE DB Command OutputCommand ExecutionAParameter information.  Matches OLE DB's DBPARAMFLAGSENUM values.Excel SourceExcel Source OutputExcel Source Error Output
Conversion<The mapping from parameters in the SQL command to variables.Excel DestinationExcel Destination InputExcel Destination Error OutputInsert�If true, values supplied for identity columns will be inserted at the destination.  If false, values for identity columns will be auto-generated at the destination.  Applies only if fast load is turned on.�If true, columns containing null will have null inserted at the destination.  If false, columns containing null will have their default values inserted at the destination.  Applies only if fast load is turned on.USpecifies options to be used with fast load.  Applies only if fast load is turned on.�Specifies when commits are issued during data insertion.  A value of 0 specifies that one commit will be issued at the end of data insertion.  Applies only if fast load is turned on.OLE DB CommandConnects and extracts data from worksheets or named ranges in Microsoft Excel workbooks. Choose one of four access modes: table or view, table or view specified in a variable, the results of a Transact-SQL statement, or the results of a Transact-SQL statement stored in a variable.�Loads data into worksheets or named ranges in Microsoft Excel workbooks. For example, update departmental reports that are currently maintained in Excel worksheets. Note: a worksheet or range is the equivalent of a table or view.There was no error.$Deferred accessor validation failed.EThe value could not be converted because of a potential loss of data.&A null value was sent to the provider.RText was truncated or one or more characters had no match in the target code page.6A signed value was provided for an unsigned data type.GConversion failed because the data value overflowed the specified type.DBSTATUS_E_CANTCREATEDBSTATUS_UNAVAILABLE6User does not have permission to write to this column.<The value violated the integrity constraints for the column.:The value violated the schema's constraint for the column.5The status value given to the provider was not valid.JThe provider used the default value for an input parameter in a procedure.The status value is unknown._The number of seconds before a command times out.  A value of 0 indicates an infinite time-out.@Specifies the name of the database object used to open a rowset.[Specifies the variable that contains the name of the database object used to open a rowset.The SQL command to be executed./Specifies the mode used to access the database.:The OLE DB runtime connection used to access the database.
OpenRowsetOpenRowset From VariableSQL CommandOpenRowset Using FastLoad'OpenRowset Using FastLoad From Variable:The variable that contains the SQL command to be executed.SQL Command From VariableeSpecifies the column code page to use when code page information is unavailable from the data source.TForces the use of the DefaultCodePage property value when describing character data.There was no error.$Deferred accessor validation failed.EThe value could not be converted because of a potential loss of data.$A null value was sent to the driver.RText was truncated or one or more characters had no match in the target code page.6A signed value was provided for an unsigned data type.GConversion failed because the data value overflowed the specified type.DBSTATUS_E_CANTCREATEDBSTATUS_UNAVAILABLE6User does not have permission to write to this column.<The value violated the integrity constraints for the column.:The value violated the schema's constraint for the column.3The status value given to the driver was not valid.HThe driver used the default value for an input parameter in a procedure.The status value is unknown.$The name of the table to be fetched.The SQL command to be executed.-Specify the number of rows fetched per batch.4Determines the chunk size allocation for LOB columns�Determines how the component exposes CHAR columns. If False, CHAR columns are exposed as (multi-)byte strings (DT_STR), if True, CHAR columns are exposed as wide strings (DT_WSTR)bThe number of seconds to wait for an SQL statement to execute before returning to the application.TThe code page to use when code page information is unavailable from the data source.%The mode used to access the database."The method used for fetching data..The ODBC C data type for binding numeric data.�Determines how the component would bind columns of (multi)-byte string types (i.e., SQL_CHAR, SQL_VARCHAR, SQL_LONGVARCHAR). The options are: Unicode (SQL_C_WCHAR) and ANSI (SQL_C_CHAR).8The ODBC runtime connection used to access the database.
Table NameSQL Command7The number of rows inserted under the same transaction.PThe method used for inserting data. Possible values are 'Batch' or 'Row by row'.There was no error.$Deferred accessor validation failed.EThe value could not be converted because of a potential loss of data.$A null value was sent to the driver.RText was truncated or one or more characters had no match in the target code page.6A signed value was provided for an unsigned data type.GConversion failed because the data value overflowed the specified type.DBSTATUS_E_CANTCREATEDBSTATUS_UNAVAILABLE6User does not have permission to write to this column.<The value violated the integrity constraints for the column.:The value violated the schema's constraint for the column.3The status value given to the driver was not valid.HThe driver used the default value for an input parameter in a procedure.The status value is unknown.$The name of the table to be fetched.The SQL command to be executed.-Specify the number of rows fetched per batch.4Determines the chunk size allocation for LOB columns�Determines how the component exposes CHAR columns. If False, CHAR columns are exposed as (multi-)byte strings (DT_STR), if True, CHAR columns are exposed as wide strings (DT_WSTR)bThe number of seconds to wait for an SQL statement to execute before returning to the application.TThe code page to use when code page information is unavailable from the data source.%The mode used to access the database."The method used for fetching data.0The ORACLE C data type for binding numeric data.�Determines how the component would bind columns of (multi)-byte string types (i.e., SQL_CHAR, SQL_VARCHAR, SQL_LONGVARCHAR). The options are: Unicode (SQL_C_WCHAR) and ANSI (SQL_C_CHAR).:The ORACLE runtime connection used to access the database.
Table NameSQL Command7The number of rows inserted under the same transaction.PThe method used for inserting data. Possible values are 'Batch' or 'Row by row'.Apply conventional bulk load.Maximum errors allowed.MDisables target table logging. Used to minimize the amount of redo generated.-Enables parallel loading of the target table.)Table sub-name (Partition, SubPartition).Transfer data buffer size (KB).Number of pre-fetched rows.Flat File DestinationFlat File Destination Input"Flat File Destination Error OutputWrites to a text file. You can specify fixed widths for data columns, use delimiters to identify columns and rows, or use both techniques for a mixed format. Choose to overwrite or append the destination file if it already exists. Use the Flat File connection manager to access the file.Flat File SourceFlat File Source OutputFlat File Source Error Output$Flat File Source Error Output Column
ConversionNIndicates whether the column uses the faster, locale-neutral parsing routines.�Reads a text file. Specify fixed widths for data columns or use delimiters to identify columns and rows. Use both techniques for files having a mixed format./Indicates whether the data is in binary format.:Specifies whether zero-length columns are treated as null.�Specifies the name of an output column containing the file name. If no name is specified, no output column containing the file name will be generated.LSpecifies whether the data will overwrite or append to the destination file.OSpecifies the text to write to the destination file before any data is written.�When text qualifier is enabled, specifies whether the text qualifier in the data written to the destination file will be escaped or not.Conditional SplitRoutes data rows to different outputs depending on the content of the data. Use conditions (SSIS expressions) to specify which rows are routed. For example, separate records that need to be cleaned from those that are ready to be loaded or route only a subset of records. Conditional Split Default Output6Default Output of the Conditional Split TransformationConditional Split Error Output4Error Output of the Conditional Split TransformationConditional Split Input-Input to the Conditional Split TransformationConditional Split Output %d1Output %d of the Conditional Split TransformationcSpecifies the expression. This expression version uses lineage identifiers instead of column names.\Specifies the friendly version of the expression. This expression version uses column names.�Specifies the position of the condition in the list of conditions that the transformation evaluates. The evaluation order is from the lowest to the highest value.Computation�Runs an SQL statement for each row in a data flow. For example, call a 'new employee setup' stored procedure for each row in the 'new employees' table. Note: running an SQL statement for each row of a large data flow may take a long time.OLE DB Command InputCache TransformWrites data to a cache (a .caw file) with the Cache Connection Manager. Reuse the cache between multiple Lookup transforms to reduce database load. Share the cache to reduce memory usage. Note: cache data is stored in clear text and not recommended for sensitive data.Cache InputCache Output(The connection used to access the cache.2The cache column associated with the input column.ODBC Source�Extracts data from an Open Database Connectivity (ODBC) compliant database. Extract from a database table or view, or use a SQL command.ODBC DestinationlLoads data into an Open Database Connectivity (ODBC) compliant database. Load into a database table or view.
ORACLE SourcejExtracts data from ORACLE compliant database. Extract from a database table or view, or use a SQL command.ODBORACLEC DestinationNLoads data into ORACLE compliant database. Load into a database table or view.Balanced Data DistributorAEvaluates and directs rows in a dataset to multiple destinations.#Balanced Data Distributor Output %dBalanced Data Distributor Input6Input to the Balanced Data	 Distributor Transformation
Conversion�Extracts data from an ODBC-compliant relational database. Extract from a database table or view, or use a SQL command. For example, extract data from tables in SQL Server databases.ODBC SourceODBC Source OutputODBC Source Error OutputODBC DestinationODBC Destination InputODBC Destination Error OutputInsert
Loads data into an ODBC-compliant relational database, such as SQL Server. Many types of databases are ODBC-compliant. With minor reconfiguration, this destination can be used before and after your database is migrated from one ODBC-compliant technology to another. 
Oracle SourceOracle Source OutputOracle Source Error Output
Conversion&Extracts data from an Oracle database.Oracle DestinationOracle Destination InputOracle Destination Error OutputInsert#Loads data into an Oracle database.�4VS_VERSION_INFO������?(StringFileInfo040904B0PlatformNT�ILegalTrademarksMicrosoft SQL Server is a registered trademark of Microsoft Corporation. CommentsSQL&GoldenBitsTrueLCompanyNameMicrosoft Corporation�8FileDescriptionData Transformation Services Execution Engine Resourcesx,FileVersion2019.0150.2000.05 ((SQLServer).190924-2033)8InternalNameDTSPipelined LegalCopyrightMicrosoft. All rights reserved.HOriginalFilenameDTSPipeline.RLLJProductNameMicrosoft SQL Server<ProductVersion15.0.2000.5DVarFileInfo$Translation	�<?xml version='1.0' encoding='UTF-8' standalone='yes'?>
<assembly xmlns='urn:schemas-microsoft-com:asm.v1' manifestVersion='1.0'>
  <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
    <security>
      <requestedPrivileges>
        <requestedExecutionLevel level='asInvoker' uiAccess='false' />
      </requestedPrivileges>
    </security>
  </trustInfo>
</assembly>
h>0�>Z	*�H��
��>K0�>G10	+0L
+�7�>0<0
+�70	���0!0	+�j�X��H?`��i��l���0��0�Ơ3,�QI�6;n,0
	*�H��
0w10	UUS10U
Washington10URedmond10U
Microsoft Corporation1!0UMicrosoft Time-Stamp PCA0
190906204002Z
201204204002Z0��10	UUS10U
Washington10URedmond10U
Microsoft Corporation1)0'U Microsoft Operations Puerto Rico1&0$UThales TSS ESN:F528-3777-8A761%0#UMicrosoft Time-Stamp Service0�"0
	*�H��
�0�
�����O�m�v�+ҳB��'a>�i��19 7�B9o�H
	wO�pOϊTH�n�������A�J���dzG�H�5�i@���W ��7ԚV��Fzc1����-@�o f���w����\�vYvw���j3�釡�A����r�_tN	���a8�~u�#��v�q�')O�.ʝ�b�@Y�Nب2�� �b�{���a��w�q�滨�4�ڏ/Ԫ�IH*�1�.g�vRAk��S,G½��[�kDRdW����	0�0U��O�f��Y1aY��3~�<�0U#0�#4��RFp
�@�v��+��5�0TUM0K0I�G�E�Chttp://crl.microsoft.com/pki/crl/products/MicrosoftTimeStampPCA.crl0X+L0J0H+0�<http://www.microsoft.com/pki/certs/MicrosoftTimeStampPCA.crt0U%0
+0
	*�H��
�$���"�*��OCX�쏅.���t��l���n`6TMM/��e�?Ɉ�ã�
�жu]�)`���O�v��|g�ib?�&��9
���a�_��@�3����Rv�F(�|��z���>�b������>u�^�5�9�qEa�|����%&�N!����|.5���F�����4�\2��E�N�
0,��!�8�sy�8��j��/_D��<�+m�<[l Fu5�x��Lr'r���|Vw���
Cd��z6�0��0�ߠ3�|��`����0
	*�H��
0y10	UUS10U
Washington10URedmond10U
Microsoft Corporation1#0!UMicrosoft Code Signing PCA0
190502212033Z
200502212033Z0t10	UUS10U
Washington10URedmond10U
Microsoft Corporation10UMicrosoft Corporation0�"0
	*�H��
�0�
����k*)�i퐭��՗+��IJ���E �Z:p�v�@�'�����Z3qʱ����^(��D�P��sY�Ep;��o�\ �(1���䱚GM<�J��f�,�U�muј^�;�F́X\��N��W���
%�%W�2�J~A��u}�ߋ�Nf
����L��,���i�HqxWw��� ����m.������:d�4���̑����yϼ̕}|�@�n���L`�!mW5�(ӭ
�܉����{0�w0U%0
+0Up���mu���*:'YG�0TUM0K�I0G1-0+U$Microsoft Ireland Operations Limited10U
229803+4541060U#0��p�R4���'���ؽ�mGN0bU[0Y0W�U�S�Qhttp://crl.microsoft.com/pki/crl/products/Microsoft%20Code%20Signing%20PCA(2).crl0f+Z0X0V+0�Jhttp://www.microsoft.com/pki/certs/Microsoft%20Code%20Signing%20PCA(2).crt0
	*�H��
�N��{�b~�=�G@������Q��Q��F�!e^bm�5�Z�꼣�h����9qp�g�&��}ZJI�δ7�909��yD�oTRIn>����X�~�G�L"x���3'��
L8*��D�e��h��JY+{p���X�6����-'����ۦ����b�شx�/.�9��5t��"5>��'���`��$yN`҄�u��J
E�:|�M�:�g�Q��#'$غ�Ȳ�1��
���A\e�0��0���
a5E?0
	*�H��
0_10
	�&���,dcom10
	�&���,d	microsoft1-0+U$Microsoft Root Certificate Authority0
180920174200Z
210509232813Z0y10	UUS10U
Washington10URedmond10U
Microsoft Corporation1#0!UMicrosoft Code Signing PCA0�"0
	*�H��
�0�
��������F��fɧ]
~�=���1��ԅ���2����A��7��|W����Z`�1
a�"�|V4b�^Z�.�@*�=�aD؎�6Y�c���������_+�P_�	Ґ頷,DA�eR7,�����,���c�Q����8^|�����)@ߔv*�e��m�'�§���j�Ƨ�f�hw�^��	�R�sQ�*�Q#�I�����P^�Oc?�/���M+q�Y('M��^0�Z0U�0�0U�p�R4���'���ؽ�mGN0U�0	+�70#	+�7<����Up�iB���8��W0	+�7
SubCA0U#0���`@V'��%�*�
S�Y�0PUI0G0E�C�A�?http://crl.microsoft.com/pki/crl/products/microsoftrootcert.crl0T+H0F0D+0�8http://www.microsoft.com/pki/certs/MicrosoftRootCert.crt0
	*�H��
�={�.�6�B�UO�����9
������"˅yC�����5�˅�->M�Q�$����c������
m.U5�,yaƶM�V&��e��.�<He��D���4
ީ�c����d�U�����e~���0���;��u�]π��q�B��ғ��T�|!�싼�~�)	P�+:�=��W��8��ߚ8Ё�'��y;4���"�e�La<��.2h�)���v�߆)�
㴳�L�gZT�Y'��HBn1L�L��h�#J�T㔕l����(�YD���򨅳=�*�7��=7�8�*G�gY�-�z�w��JI*t7��/��C-�2���LI�����G辞�1d������uI{db��~��g\AG�N6-�7�k�?R����"%�9[�Ez��������d�΋��}��mW[*{ol~�C�
�]�gԋ���Nz���߽4����gn�d��L�3��QpU�\}�ݧU��ǣ�Fz��D��c�}�"u0�0��
ah40
	*�H��
0_10
	�&���,dcom10
	�&���,d	microsoft1-0+U$Microsoft Root Certificate Authority0
070403125309Z
210403130309Z0w10	UUS10U
Washington10URedmond10U
Microsoft Corporation1!0UMicrosoft Time-Stamp PCA0�"0
	*�H��
�0�
���l���H�*|k.���řQ#P���N$�ѧ�L�t0 ��T�I�B���\0&�k�{�������3$����g,�{���N���#Sʍ�E�	�^�аB���l�g�(�FS����(F.'��OUR�,�j��0n��Y0��~֗��!�~����SmR�E��HJGf')f�����J����G	����Ux��P�0F��#F��-T��xA1w̛�;���a�
������"KT���� ���0��0U�0�0U#4��RFp
�@�v��+��5�0U�0	+�70��U#��0�����`@V'��%�*�
S�Y䤡c�a0_10
	�&���,dcom10
	�&���,d	microsoft1-0+U$Microsoft Root Certificate Authority�y��J���LsX�.e0PUI0G0E�C�A�?http://crl.microsoft.com/pki/crl/products/microsoftrootcert.crl0T+H0F0D+0�8http://www.microsoft.com/pki/certs/MicrosoftRootCert.crt0U%0
+0
	*�H��
����\D6�鴭w��yQM�.tq[m�ξ{��.���mb�W�N P��z@�W��X�����x���;�⾏
ȍH��S�O����jՀ���/��*$"p����x�.
��p�N��k
r���9�g-�Q�b\?�� ?�2��1Jv1=+r��p=��
��
a��׸�N��5+^���}B{�Sr!�0ʻxe\[
����<�fX�-/���QV�p�l�9L�y��zb))4wb�=m���,��C���}�㤨t5�N�q#&v<R�qZFC����MgȟB��y!��0f�B^Ȟ"!�C�+��H�,��2�L&*H$_Z�V�ޙ��:o��'t�v!����'T��j=��va�n����G�o��2l���QȌ#�f���d=-�Y���!嵰�c�xT�A.P����x����uْ�u"/2�n�[�%����!��Oɚ�I�eo�}�i��n꧗�1)��[\�M��/�?.Uk���1�(90�(50��0y10	UUS10U
Washington10URedmond10U
Microsoft Corporation1#0!UMicrosoft Code Signing PCA3�|��`����0	+���0	*�H��
	1
+�70
+�710
+�70#	*�H��
	1(�Ū��w�Ɠ��t���n;�0N
+�71@0>� �SQL Server 2019��http://www.microsoft.com0
	*�H��
�s�D�8vԹ-6��!UJK��J,�*�z{r e-D�f��X��ׯ4�K������ضy�����_b�ޞ}�t��R٦D_�jԧ���-�eh�	�
��_�t�U
J�3r���8�q�'|���n�%��	�}�����w�1o��9����>����J�&�%�N��0��r'L]R	�&����}[��{��4/̂AS�T�H�,ӆ����,�m��N{�0�;�Rp��[[�=���ڪ��0��%�0�$	*�H��
	1�0�0��0w10	UUS10U
Washington10URedmond10U
Microsoft Corporation1!0UMicrosoft Time-Stamp PCA3,�QI�6;n,0	+�]0	*�H��
	1	*�H��
0	*�H��
	1
190924204415Z0#	*�H��
	1�[!U��~��qA�ڥ𺝬0
	*�H��
�'6�=��B�q��C�#���1�W�fd졍-\�2D����V�� ��
k�\؃.�y^m�D��ֲ.��i��r��0ҩ�lY��t��%Xn�+U���08��U�C�t��&��t���Ea��	��*��EUs���FZ��OG?�P��������͔�
�l�;�ݩ��v�}KO���&�^?��`�$��w�]a]��ԋ��?��XSh�i9��R-��}i�E���E�:sw��Cߦ
0�#�
+�71�#�0�#�	*�H��
��#}0�#y10
	`�He0\
+�7�N0L0
+�70	���010
	`�He �P�)딺�30$	�oZ��4
T|v���T��<ŝ���
�0�0��3R�@�PV�u�R0
	*�H��
0~10	UUS10U
Washington10URedmond10U
Microsoft Corporation1(0&UMicrosoft Code Signing PCA 20110
190502213746Z
200502213746Z0t10	UUS10U
Washington10URedmond10U
Microsoft Corporation10UMicrosoft Corporation0�"0
	*�H��
�0�
���������t�����ǔߖ�6�+�xL���5���oK��w�U��buAN��[�V7V�p�Cn�3(A6��~��/�k����>��\�6��ɳ����mJ�
�X�%y�D�����*��Or�}Y�Gڗ}tP����M�4dc�睘�*�!�U��}�W����P��e���.����lA ��+4�Ih��z%�x�tr�
����������c� �����i��Zai^���76q�3ހ����0�~0U%0
+�7L+0U��[�e�'f��5kE�0TUM0K�I0G1-0+U$Microsoft Ireland Operations Limited10U
230012+4541360U#0�Hnd�Pӂ�77"�m��u�0TUM0K0I�G�E�Chttp://www.microsoft.com/pkiops/crl/MicCodSigPCA2011_2011-07-08.crl0a+U0S0Q+0�Ehttp://www.microsoft.com/pkiops/certs/MicCodSigPCA2011_2011-07-08.crt0U�00
	*�H��
���D懾��,V�(�-����6@��m�z� 65�W��Q`[�F�y8��t�ھ$�&��C��=�d�R�m���(ߜ����S�]��!��h2]g����"���S��E��v���5@��7�8��t�=0|3e�?�O��%�vS�Y�������ޭ��b�����+O�yDT
�ͅC�8xQ#�+7&�T��",ۮ8��P���h�(�T@�`���T�l������T����G6~�ܷ�f��D�k�IB���	��m������r����aݽ��G>ƚ읶��2>�Ҁ\�����]��r!���K��Q!��(�V�&sLDt�>��QZ�M����W5X��)l�}F�G�M3�s4��O��8v�vT���:�e:����*Ɲ�N�
���hP���z?别2��fwǴ��#cU�b���ӧx�\Ȇ޽�e,��Ul*!����43=+�
�m����
 ��ސ��n���Iy/钒>
�0�z0�b�
a��0
	*�H��
0��10	UUS10U
Washington10URedmond10U
Microsoft Corporation1200U)Microsoft Root Certificate Authority 20110
110708205909Z
260708210909Z0~10	UUS10U
Washington10URedmond10U
Microsoft Corporation1(0&UMicrosoft Code Signing PCA 20110�"0
	*�H��
�0�
���r.��n��M4��X!�B*k�ZP��8I���7k���8™��9�1BӉ
yd�~�`$l��I��h^�ߛS�
,ï٩+�z	�זY�`�fv�2R&/�PϳD�]��.u�h��m�:\��F��8d�nd5x��c-�@����
�\y�I)�
��w��=��+ZMV(�zr�:�I^���C���{�{��>��]�O��/Y�;-�3Xŷ>�2ӳ=����8~ҝ�,�NV�!5%�9ndS���#������_��R������!t%_(��'(8%�9J6�|���#��f�aj�(I�_��%]�!K>RĵW?$��z[/�#�p]QFw��Ἤ�_����UK�9��#I��D|E��rz�r�$߿F����Wۃ��MI0�����[���ݰf��{���K�I)�(��}g��bx_�/��W�\�w((��m�(�,�@7O��D��	L�ԥC/t��n�x X,]`��>O3�ڰ�ޞN�F�l���ֈ����0��0	+�70UHnd�Pӂ�77"�m��u�0	+�7
SubCA0U�0U�0�0U#0�r-:1�C�N���1�#�40ZUS0Q0O�M�K�Ihttp://crl.microsoft.com/pki/crl/products/MicRooCerAut2011_2011_03_22.crl0^+R0P0N+0�Bhttp://www.microsoft.com/pki/certs/MicRooCerAut2011_2011_03_22.crt0��U ��0��0��	+�7.0��0?+3http://www.microsoft.com/pkiops/docs/primarycps.htm0@+042 Legal_policy_statement. 0
	*�H��
�g򆥘�Ty.��tg"���c��B�}�y��e_.,>��r�m���?�
�;�G���i�c�"5��]e��}�FPU��|�K�<B�Rа�k�>.)��=�Եw�9��d�D=xz#�}��t����&F*Š�����h�h.��*?*kXIc	inZ�����F+�;н5�n%��'������(��M=����2���-�[��I9
�
�ƭ�����QEX82q'��'��,: i��Y�hno��t�@�*(>�?Mf��M��o}E�v�@*e��]Rb��c6��������'�ͥ���]cA�[��>w��u�aa������A�(��e/��\��֡F�Y�	KH���)ű��?Ew5�Ң��zz"�H���G#�����k�K���ux��7�yK�ֳ#hu������i;��0�L���umc�=ѝVNO��W"�x2!z�A
�?��̤]����WqVNHE�Bɛv[
�Hk�y���mmj�Rs
zP�1�z0�v0��0~10	UUS10U
Washington10URedmond10U
Microsoft Corporation1(0&UMicrosoft Code Signing PCA 20113R�@�PV�u�R0
	`�He���0
*�H��
	10	*�H��
	1
+�70
+�710
+�70/	*�H��
	1" ѻh;�IĬb�LJ�S+��{D�
R��0N
+�71@0>� �SQL Server 2019��http://www.microsoft.com0
	*�H��
�Z�hjO:����>����u���&�麰��9�*������WJ����X��H�����G���F���Y�[f�#G
v��B��+|�wy�����0�E�뚺=��=��6��,1��<ͨ���'K׹��3�3@_+��{}{�%��}�����n�!席�������:0h�*>P���y<Q�M%�4�e|�gIz{@Dq��'���'W��wC��Ar$�<�JQf�g��@��2Bm�%
�d^%�b+u���0��
+�71��0��	*�H��
���0��10
	`�He0�Q*�H��
	��@�<0�8
+�Y
010
	`�He �f�aJ����'6�x$��@��Q��f�$�G]^��8�20190924204417.016Z0����0��10	UUS10	UWA10URedmond10U
Microsoft Corporation1-0+U$Microsoft Ireland Operations Limited1&0$UThales TSS ESN:179E-4BB0-82461%0#UMicrosoft Time-Stamp service��<0��0�٠3۪m�"4%�0
	*�H��
0|10	UUS10U
Washington10URedmond10U
Microsoft Corporation1&0$UMicrosoft Time-Stamp PCA 20100
180823202653Z
191123202653Z0��10	UUS10	UWA10URedmond10U
Microsoft Corporation1-0+U$Microsoft Ireland Operations Limited1&0$UThales TSS ESN:179E-4BB0-82461%0#UMicrosoft Time-Stamp service0�"0
	*�H��
�0�
������$�K6n�;�OT\�S
��0Ϊ 0ޖ��&
u0�	�%
�M�R��t����	a��)���V	���������ݡ�cfc̳?v�H���j��`&,o�2~�n�}"��(&�gw;��eb _���T?k=#�)e!�9�E��u��"�F�Q�s���-<���5�߷Ԫ�8�P��?�Da��:�ֱ
�㾁b����\C��o"�Y]M,#�
�4/�z6�-�ods1��ӳ%/�M��0�0U����\s�Ͷ�����*CH0U#0��c:\�1��C{|F�3hZ�mU0VUO0M0K�I�G�Ehttp://crl.microsoft.com/pki/crl/products/MicTimStaPCA_2010-07-01.crl0Z+N0L0J+0�>http://www.microsoft.com/pki/certs/MicTimStaPCA_2010-07-01.crt0U�00U%0
+0
	*�H��
�Bš��`~"G�$違�^nWcQ��G엓Y1LXi ���O���P2���8@��q�Vv(��C�f�d}Z���,�������ٹ�0�z;�F����k/�`���vȂ�Lu���=�Et@�������B��O+�G8����,����=�ɍI>��ѯ�1W��3�—Fg����>ݭ�2�LF�ld�ȣq��m��Df�\��_����R�XX��{���Bs�Q�J�p�:ZTd=!�0�q0�Y�
a	�*0
	*�H��
0��10	UUS10U
Washington10URedmond10U
Microsoft Corporation1200U)Microsoft Root Certificate Authority 20100
100701213655Z
250701214655Z0|10	UUS10U
Washington10URedmond10U
Microsoft Corporation1&0$UMicrosoft Time-Stamp PCA 20100�"0
	*�H��
�0�
��
�w�: ����i�ktTե
����|hK,_���a�v�>f+[�S'1A��	�|a0Y�0D�`��TC�M�8�Bݓ��s0W&��E
��G�Ϳ$`2X`F�XG�2�tag�_�T�ϓL��Ħ]�an(������a�F�'$gr!��Kd�Pb�]�w=�Wu���BM@Q��>g�f��D~��K�n�ʱ��z*
1��N���5�x���<�/D����d��� 	�x����D]^�O"M���0��0	+�70U�c:\�1��C{|F�3hZ�mU0	+�7
SubCA0U�0U�0�0U#0��Vˏ�\bh�=��[�Κ�0VUO0M0K�I�G�Ehttp://crl.microsoft.com/pki/crl/products/MicRooCerAut_2010-06-23.crl0Z+N0L0J+0�>http://www.microsoft.com/pki/certs/MicRooCerAut_2010-06-23.crt0��U ���0��0��	+�7.0��0=+1http://www.microsoft.com/PKI/docs/CPS/default.htm0@+042 Legal_Policy_Statement. 0
	*�H��
��Q
���?�q=���!o���1����Wm0���f�j���x�Ǩ�%��kTW+Q�D��S���`�v�@@A��\�\�^5�$VKt�Bȯ���7}Z?�yJ�R�8�/y�e٩�k����zoK0D$"<�����Y)����p�2J'�U�/���3�b_��W@��Ιf���jb��J�&�9�Jqc�{!mÜ��<�}�j�x�m���8ؔ�ƥ
������B�����"8 �%�d��~cY%z.9Wv�q����a�˚�G����ͧ��}���;q	�]t�"aڰPo�����1��:�eGx�H���r~akow��˧	�9����؂�r�����*T9�[��
U�z�s;�-��3.)��/T'!�ȬN�(���ۖ��B���AM�*��f0ӻt2K�c{���/�!�Y���D<�Pqס��U�i�W���0����M�
�]O�8/��X.�
�P5��	'ճ~5��6��_��t�I���0�70����Ф��0��10	UUS10	UWA10URedmond10U
Microsoft Corporation1-0+U$Microsoft Ireland Operations Limited1&0$UThales TSS ESN:179E-4BB0-82461%0#UMicrosoft Time-Stamp service�#
0+[�)N���p<$OjD[�o��0���0���~0|10	UUS10U
Washington10URedmond10U
Microsoft Corporation1&0$UMicrosoft Time-Stamp PCA 20100
	*�H��
�4�L0"20190925004356Z20190926004356Z0w0=
+�Y
1/0-0
�4�L0
��0l0
�6�06
+�Y
1(0&0
+�Y
�
0� �
0��0
	*�H��
��qVЂ�i]�FB����������I��k��>��s�>�tB�їDā���^oec[$v�!,�ҽG����i�P�#�=I��:b�>c-ʦ�8u{ւ��:����zd鿪3��,��$��q^�=p�%�Eշ1�
0�	0��0|10	UUS10U
Washington10URedmond10U
Microsoft Corporation1&0$UMicrosoft Time-Stamp PCA 20103۪m�"4%�0
	`�He��J0	*�H��
	1
*�H��
	0/	*�H��
	1" 2: ��B)��q���s������Kv�
0��*�H��
	/1��0��0��0�� S�Ә-"�j#%c��}G�� ������!���0��0���~0|10	UUS10U
Washington10URedmond10U
Microsoft Corporation1&0$UMicrosoft Time-Stamp PCA 20103۪m�"4%�0" 0ƺ9e�s	���l���}I6��՛Ɣ9!{50
	*�H��
��j�#ݳ3r:MO�0p���Ȑ��;fp��v^+�� Qw�qͰ�������("�m�ԏt3�
#��R�2���5�l��Fb27���Bd$�1���}5}��ڇ�g�="��	%�!
%7�ڰ��M��Ѣ#�b���_�5�Cy+����#$3�L�������6t"��)��w�ک>w�5ɖWW�uiN�s/M���j8_�	oq�@��8��o0�
�9H��fw��}��V1o�z�iH