@@ -302,7 +302,7 @@ static auto registry0 = torch::RegisterOperators().op(
302302 []() {
303303 })
304304).op(
305- " _prim::ListConstruct.Tensor " ,
305+ " _prim::ListConstruct.tensor " ,
306306 torch::RegisterOperators::options ().catchAllKernel(
307307 []() {
308308 })
@@ -311,143 +311,6 @@ static auto registry0 = torch::RegisterOperators().op(
311311 torch::RegisterOperators::options ().catchAllKernel(
312312 []() {
313313 })
314- // Pytext operators
315- ).op(
316- " _aten::embedding(Tensor weight, Tensor indices, int padding_idx=-1, bool scale_grad_by_freq=False, bool sparse=False) -> Tensor" ,
317- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
318- [](c10::OperatorKernel* kernel, Stack* stack) {
319- constexpr int N = 5 ;
320- auto result_ = at::embedding (
321- (std::move (peek (*stack, 0 , N))).toTensor (),
322- (std::move (peek (*stack, 1 , N))).toTensor (),
323- (std::move (peek (*stack, 2 , N))).toInt (),
324- (std::move (peek (*stack, 3 , N))).toBool (),
325- (std::move (peek (*stack, 4 , N))).toBool ()
326- );
327- drop (*stack, N);
328- pack (*stack, std::move (result_));
329- })
330- ).op(
331- " _aten::dropout(Tensor input, float p, bool train) -> Tensor" ,
332- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
333- [](c10::OperatorKernel* kernel, Stack* stack) {
334- auto result_ = at::dropout (
335- (std::move (peek (*stack, 0 , 3 ))).toTensor (),
336- (std::move (peek (*stack, 1 , 3 ))).toDouble (),
337- (std::move (peek (*stack, 2 , 3 ))).toBool ()
338- );
339- drop (*stack, 3 );
340- pack (*stack, std::move (result_));
341- })
342- ).op(
343- " _aten::permute(Tensor(a) self, int[] dims) -> Tensor(a)" ,
344- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
345- [](c10::OperatorKernel* kernel, Stack* stack) {
346- auto result_ = ((std::move (peek (*stack, 0 , 2 ))).toTensor ()).permute (
347- (std::move (peek (*stack, 1 , 2 ))).toIntListRef ()
348- );
349- drop (*stack, 2 );
350- pack (*stack, std::move (result_));
351- }).aliasAnalysis(c10::AliasAnalysisKind::FROM_SCHEMA)
352- ).op(
353- " _aten::matmul(Tensor self, Tensor other) -> Tensor" ,
354- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
355- [](c10::OperatorKernel* kernel, Stack* stack) {
356- auto result_ = at::matmul (
357- (std::move (peek (*stack, 0 , 2 ))).toTensor (),
358- (std::move (peek (*stack, 1 , 2 ))).toTensor ()
359- );
360- drop (*stack, 2 );
361- pack (*stack, std::move (result_));
362- })
363- ).op(
364- " _aten::mul.Tensor(Tensor self, Tensor other) -> Tensor" ,
365- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
366- [](c10::OperatorKernel* kernel, Stack* stack) {
367- auto result_ = at::mul (
368- (std::move (peek (*stack, 0 , 2 ))).toTensor (),
369- (std::move (peek (*stack, 1 , 2 ))).toTensor ()
370- );
371- drop (*stack, 2 );
372- pack (*stack, std::move (result_));
373- })
374- ).op(
375- " _aten::tanh(Tensor self) -> Tensor" ,
376- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
377- [](c10::OperatorKernel* kernel, Stack* stack) {
378- auto result_ = at::tanh (
379- (std::move (peek (*stack, 0 , 1 ))).toTensor ()
380- );
381- drop (*stack, 1 );
382- pack (*stack, std::move (result_));
383- })
384- ).op(
385- " _aten::max.dim(Tensor self, int dim, bool keepdim=False) -> (Tensor values, Tensor indices)" ,
386- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
387- [](c10::OperatorKernel* kernel, Stack* stack) {
388- auto result_ = at::max (
389- (std::move (peek (*stack, 0 , 3 ))).toTensor (),
390- (std::move (peek (*stack, 1 , 3 ))).toInt (),
391- (std::move (peek (*stack, 2 , 3 ))).toBool ()
392- );
393- drop (*stack, 3 );
394- pack (*stack, std::move (result_));
395- })
396- ).op(
397- " _aten::cat(Tensor[] tensors, int dim=0) -> Tensor" ,
398- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
399- [](c10::OperatorKernel* kernel, Stack* stack) {
400- auto result_ = at::cat (
401- (std::move (peek (*stack, 0 , 2 ))).toTensorListRef (),
402- (std::move (peek (*stack, 1 , 2 ))).toInt ()
403- );
404- drop (*stack, 2 );
405- pack (*stack, std::move (result_));
406- })
407- ).op(
408- " _aten::__is__(t1 self, t2 obj) -> bool" ,
409- torch::RegisterOperators::options ().catchAllKernel(
410- [](c10::OperatorKernel* kernel, Stack* stack) {
411- c10::IValue self, obj;
412- pop (*stack, self, obj);
413- push (*stack, self.isSameIdentity (obj));
414- })
415- ).op(
416- " _aten::log_softmax.int(Tensor self, int dim, ScalarType? dtype=None) -> Tensor" ,
417- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
418- [](c10::OperatorKernel* kernel, Stack* stack) {
419- auto result_ = at::log_softmax (
420- (std::move (peek (*stack, 0 , 3 ))).toTensor (),
421- (std::move (peek (*stack, 1 , 3 ))).toInt (),
422- (std::move (peek (*stack, 2 , 3 ))).toOptional <c10::ScalarType>()
423- );
424- drop (*stack, 3 );
425- pack (*stack, std::move (result_));
426- })
427- ).op(
428- " _aten::softmax.int(Tensor self, int dim, ScalarType? dtype=None) -> Tensor" ,
429- torch::RegisterOperators::options ().kernel(c10::TensorTypeId::CPUTensorId,
430- [](c10::OperatorKernel* kernel, Stack* stack) {
431- auto result_ = at::softmax (
432- (std::move (peek (*stack, 0 , 3 ))).toTensor (),
433- (std::move (peek (*stack, 1 , 3 ))).toInt (),
434- (std::move (peek (*stack, 2 , 3 ))).toOptional <c10::ScalarType>()
435- );
436- drop (*stack, 3 );
437- pack (*stack, std::move (result_));
438- })
439- ).op(
440- " _aten::warn() -> void" ,
441- torch::RegisterOperators::options ().catchAllKernel(
442- [](c10::OperatorKernel* kernel, Stack* stack) {
443- drop (*stack, 1 );
444- pop (*stack);
445- })
446- ).op(
447- " _prim::unchecked_cast" ,
448- torch::RegisterOperators::options ().catchAllKernel(
449- []() {
450- })
451314).op(
452315 " _prim::TupleConstruct" ,
453316 torch::RegisterOperators::options ().catchAllKernel(
0 commit comments