Sketch2FullStack: Generating Skeleton Code of Full Stack Website and Application from Sketch using Deep Learning and Computer Vision

  title={Sketch2FullStack: Generating Skeleton Code of Full Stack Website and Application from Sketch using Deep Learning and Computer Vision},
  author={Somoy Subandhu Barua and Imam Mohammad Zulkarnain and Abhishek Roy and Md. Golam Rabiul Alam and Md. Zia Uddin},
—For a full-stack web or app development, it requires a software firm or more specifically a team of experienced developers to contribute a large portion of their time and resources to design the website and then convert it to code. As a result, the efficiency of the development team is significantly reduced when it comes to converting UI wireframes and database schemas into an actual working system. It would save valuable resources and fasten the overall workflow if the clients or developers can… 



Sketch2code: Generating a website from a paper mockup

This dissertation presents two approaches which automates this process of developing boiler plate user interface code, one using classical computer vision techniques, and another using a novel application of deep semantic segmentation networks.

Sketch2Code: Transformation of Sketches to UI in Real-time Using Deep Neural Network

A novel approach wherein a Deep Neural Network is employed that is trained on a custom database of such sketches to detect UI elements in the input sketch to create a UI prototype for multiple platforms with single training.

Machine Learning-Based Prototyping of Graphical User Interfaces for Mobile Apps

This paper presents an approach that automates this process by enabling accurate prototyping of GUIs via three tasks: detection, classification, and assembly, and implemented this approach for Android in a system called ReDraw.

DeepDeSRT: Deep Learning for Detection and Structure Recognition of Tables in Document Images

In contrast to most existing table detection and structure recognition methods, which are applicable only to PDFs, DeepDeSRT processes document images, which makes it equally suitable for born-digital PDFs as well as even harder problems, e.g. scanned documents.

Image-based table recognition: data, model, and evaluation

The largest publicly available table recognition dataset PubTabNet is developed, containing 568k table images with corresponding structured HTML representation, and a new Tree-Edit-Distance-based Similarity (TEDS) metric for table recognition is proposed, which more appropriately captures multi-hop cell misalignment and OCR errors than the pre-established metric.

Table Detection Using Deep Learning

The proposed method works with high precision on document images with varying layouts that include documents, research papers, and magazines and beats Tesseract's state of the art table detection system by a significant margin.

EAST: An Efficient and Accurate Scene Text Detector

This work proposes a simple yet powerful pipeline that yields fast and accurate text detection in natural scenes, and significantly outperforms state-of-the-art methods in terms of both accuracy and efficiency.

Optimizing the Trade-Off between Single-Stage and Two-Stage Deep Object Detectors using Image Difficulty Prediction

  • Petru SovianyRadu Tudor Ionescu
  • Computer Science
    2018 20th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)
  • 2018
This paper proposes to use an image difficulty predictor to achieve an optimal trade-off between accuracy and speed in object detection, and shows that using image difficulty compares favorably to a random split of the images.

CascadeTabNet: An approach for end to end table detection and structure recognition from image-based documents

This paper proposes CascadeTabNet: a Cascade mask Region-based CNN High-Resolution Network (Cascade mask R-CNN HRNet) based model that detects the regions of tables and recognizes the structural body cells from the detected tables at the same time.

Speed/Accuracy Trade-Offs for Modern Convolutional Object Detectors

A unified implementation of the Faster R-CNN, R-FCN and SSD systems is presented and the speed/accuracy trade-off curve created by using alternative feature extractors and varying other critical parameters such as image size within each of these meta-architectures is traced out.